This is a special post for quick takes by Lizka. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
(This was initially meant as part of this post,[1] but while editing I thought it didn't make a lot of sense there, so I pulled it out.)
I came to CEA with a very pro-criticism attitude. My experience there reinforced those views in some ways,[2] but it also left me more attuned to the costs of criticism (or of some pro-criticism attitudes). (For instance, I used to see engaging with all criticism as virtuous, and have changed my mind on that.) My overall takes now aren’t very crisp or easily summarizable, but I figured I'd try to share some notes.
...
It’s generally good for a community’s culture to encourage criticism, but this is more complicated than I used to think.
Here’s a list of things that I believe about criticism:
Criticism or critical information can be extremely valuable. It can be hard for people to surface criticism (e.g. because they fear repercussions), which means criticism tends to be undersupplied.[3] Requiring critics to present their criticisms in specific ways will likely stifle at least some valuable criticism. It can be hard to get yourself to engage with criticism of your work or things you care about. I
I would be excited about this and have wondered for a while if we should have EA awards. This Washington post article brought the idea to my mind again:
(This was initially meant as part of this post[1], but I thought it didn't make a lot of sense there, so I pulled it out.)
“Slow-rolling mistakes” are usually much more important to identify than “point-in-time blunders,”[2] but the latter tend to be more obvious.
When we think about “mistakes”, we usually imagine replying-all when we meant to reply only to the sender, using the wrong input in an analysis, including broken hyperlinks in a piece of media, missing a deadline, etc. I tend to feel pretty horrible when I notice that I've made a mistake like this.
I now think that basically none of my mistakes of this kind — I’ll call them “Point-in-time blunders” — mattered nearly as much as other "mistakes" I've made by doing things like planning my time poorly, delaying for too long on something, setting up poor systems, or focusing on the wrong things.
This second kind of mistake — let’s use the phrase “slow-rolling mistakes” — is harder to catch; I think sometimes I'd identify them by noticing a nagging worry, or by having multiple conversations with someone who disagreed with me (and slowly changing my mind), or by seriously reflec... (read more)
After reading your post, I wasn't sure you were right about this. But after thinking about it for a few minutes, I can't come up with any serious mistakes I've made that were "point-in-time blunders".
The closest thing I can think of is when I accidentally donated $20,000 to the GiveWell Community Foundation instead of The Clear Fund (aka GiveWell), but fortunately they returned the money so it all worked out.
For context, lead poisoning seems to get ~$11-15 million per year right now, and has a huge toll. I'm really excited about this news.
Also, thanks to @ryancbriggs for pointing out that this seems like "a huge win for risky policy change global health effective altruism" and referencing this grant:
In December 2021, GiveWell (or the EA Funds Global Health and Development Fund?) gave a grant to CGD to "to support research into the effects of lead exposure on economic and educational outcomes, and run a working group that will author policy outreach documents and engage with global policymakers." In their writeup, they recorded a 10% "best case" forecast that in two years (by the end of the grant period), "The U.S. government, other international actors (e.g., bilateral and multilateral donors), and/or national ... (read more)
This is awesome! Is there a page somewhere that collates the results of a bunch of internal forecasting by the end of the grant period? I'd be interested
Reflection on my time as a Visiting Fellow at Rethink Priorities this summer
I was a Visiting Fellow at Rethink Priorities this summer. They’re hiring right now, and I have lots of thoughts on my time there, so I figured that I’d share some. I had some misconceptions coming in, and I think I would have benefited from a post like this, so I’m guessing other people might, too. Unfortunately, I don’t have time to write anything in depth for now, so a shortform will have to do.
Fair warning: this shortform is quite personal and one-sided. In particular, when I tried to think of downsides to highlight to make this post fair, few came to mind, so the post is very upsides-heavy. (Linch’s recent post has a lot more on possible negatives about working at RP.) Another disclaimer: I changed in various ways during the summer, including in terms of my preferences and priorities. I think this is good, but there’s also a good chance of some bias (I’m happy with how working at RP went because working at RP transformed me into the kind of person who’s happy with that sort of work, etc.). (See additional disclaimer at the bottom.)
First, some vague background on me, in case it’s relevant:
Thanks for writing this Lizka! I agree with many of the points in this [I was also a visiting fellow on the longtermist team this summer]. I'll throw my two cents in about my own reflections (I broadly share Lizka's experience, so here I just highlight the upsides/downsides things that especially resonated with me, or things unique to my own situation):
Vague background:
Finished BSc in PPE this June
No EA research experience and very little academic research experience
Introduced to EA in 2019
Upsides:
Work in areas that are intellectually stimulating and feel meaningful (e.g. Democracy, AI Governance).
Working with super cool people. Everyone was super friendly, and clearly supportive of our development as researchers. I also had not written an EA forum post be
Thanks a lot for writing about your experiences, Lizka and Tom! Especially the details about why you were happy with your managers was really valuable info for me.
Protip: if you find yourself with a slow computer, fix that situation asap.
Note to onlookers that we at Rethink Priorities will pay up to $2000 for people to upgrade their computers and that we view this as very important! And if you work with us for more than a year, you can keep your new computer forever.
I realize that this policy may not be a great fit for interns / fellows though, so perhaps I will think about how we can approach that.
I think we should maybe just send a new mid-end chromebook + high-end headsets with builtin mic + other computing supplies to all interns as soon as they start (or maybe before), no questions asked. Maybe consider higher end equipment for interns who are working on more compute-intensive stuff and/or if they or their managers asked for it.
For some of the intern projects (most notably on the survey team?), more computing power is needed, but since so much of RP work involves Google docs + looking stuff up fast on the internet + Slack/Google Meet comms, the primary technological bottlenecks that we should try to solve is really fast browsing/typing/videocall latency and quality, which chromebooks and headsets should be sufficient for.
(For logistical reasons I'm assuming that the easiest thing to do is to let the interns keep the chromebook and relevant accessories)
What do you think the final voting mechanism should be, and why? E.g. approval voting, ranked-choice voting, quadratic voting, etc.
Considerations might include: how well this will allocate funds based on real preferences, how understandable it is to people who are participating in the Donation Election or following it, etc.
I realize that I might be opening a can of worms, but I'm looking forward to reading any comments! I might not have time to respond.
Users will be able to "pre-vote" (to signal that they're likely to vote for some candidates, and possibly to follow posts about some candidates), for as many candidates as they want. The pre-votes are anonymous (as are final votes), but the total numbers will be shown to everyone. There will be a separate process for final voting, which will determine the three winners in the election. The three winners will receive the winnings from the Donation Election Fund, split proportionally based on the votes.
I’m a researcher on voting theory, with a focus on voting over how to divide a budget between uses. Sorry I found this post late, so probably things are already decided but I thought I’d add my thoughts. I’m going to assume approval voting as input format.
There is an important high-level decision to make first regarding the objective: do we want to pick charities with the highest support (majoritarian) or do we want to give everyone equal influence on the outcome if possible (proportionality)?
If the answer is “majoritarian”, then the simplest method makes the most sense: give all the money to the charity with the highest approval score. (This maximizes the sum of voter utilities, if you define voter utility to be the amount of money that goes to the charities a voter approves.)
If the answer is “proportionality”, my top recommendation would be to drop the idea of having only 3 winners and not impose a limit, and instead use the Nash Product rule to decide how the money is split [paper, wikipedia]. This rule has a nice interpretation where let’s say there are 100 voters, then every voter is assigned 1/100th of the budget and gets a guarantee that this part is only spent on charities ... (read more)
This definition of "voter utility" feels very different to how EAs think about charities: the definition would imply that you are indifferent between all charities that you approve of. A better definition of "voter utility" would take into account the relative worth of the charities (eg a voter might think that charitiy A is 3x better than charity B, which is 5x better than charity C).
I think since there can be multiple winners, letting people vote on the ideal distribution then averaging those distributions would be better than direct voting, since it most directly represents "how voters think the funds should be split on average" or similar, which seems like what you want to capture? And also is still very understandable I hope.
E.g. if I think 75% of the pool should go to LTFF and 20% to GiveWell, and 5% to the EA AWF, 0% to all the rest, I vote 75%/20%/5%/0%/0%/0% etc. Then, you take the average of those distributions across all voters. I guess it gets tricky if you are only paying out to the top three, but maybe you can just scale their percentage splits? IDK.
If not that or if it is annoying to implement, IMO approval voting or quadratic are probably best, but am not really sure. Ranked choice feels like it is so explicitly designed for single winner elections that it is harder to apply here.
If we're thinking of it as "ideally I'd like 75% of the money to go here, 20% here, etc" we could just give people 100 votes each and give money to the top 3?
2
abrahamrowe
Yeah definitely - that's a more elegant way.
1
harfe
This would be very similar to first-past-the-post (third-past-the-post in this case), and has many of the same drawbacks as first-past-the-post, such as lots of strategic voting. Giving a voice to people who's favorite charities are not wildly popular seems preferable (as would be the case with ranked-choice voting). The fact that you have 100 votes instead of 1 vote doesnt make much of a difference here (imagine a country where everyone has 99 clones, election systems would mostly still have the same advantages and disadvantages).
I think this could be fun. An advantage here is that voters have to think about the relative value of different charities, rather than just deciding which are better or worse. This could also be an important aspect when we want people to discuss how they plan to vote/how others should vote.
If you want to be explicit about this, you could also consider designing the user interface so that users enter these relative differences of charities directly (e.g. "I vote charity A to be 3 times as good as charity B" rather than "I assign 90 vote credits to charity A and 10 vote credits to charity B").
Note however, that due to the top-3 cutoff, putting in the true relative differences between charities might not be the optimal policy.
A technical remark: If you want only to do payouts for the top three candidates, instead of just relying on the final vote, I think it would be better to rescale the voting credits of each voter after kicking out the charity with the least votes and then repeating the process until there are only 3 charities left. This would reduce tactical voting and would respect voters more who pick unusual charities as ... (read more)
One issue that comes up with multi-winner approval voting is: suppose there are 15 longtermists and 10 global poverty people. All the longtermists approve the LTFF, MIRI, and Redwood; all the global poverty people approve the Against Malaria Foundation, GiveWell, and LEEP.
The top three vote winners are picked: they're the LTFF, with 15 votes, MIRI, with 15 votes, and Redwood, with 15 votes.
It is maybe undesirable that 40% of the people in this toy example think those charities are useless, yet 0% of money is going to charities that aren't those. (Or maybe it's not! If a coin lands heads 60% of the time; then you bet on heads 100% of the time.)
5
JWS 🔸
I'm going to stick my neck out and say that approval voting is the best option here. Why?
1. It avoids almost all of the problems with plurality voting. In non-pathological arrangments of voter preferences and candidates, it will produce the 'intuitively' correct option - see here for some fun visualisations.
2. It has EA cred, see Aaron Hamlin's interview on 80k here
3. And most importantly, it's understandable and legible - you don't need people to trust an underlying apportionment algorithm or send the flyers explaining the D'Hondt method to voters or whatever. Just vote for the options you approve of on the ballot. One person, one ballot. Most approvals wins. Simple.
I fear that EAs who are really into this sort-of thing are going to nerd-snipe the whole thing into a discussion/natural experiment about optimal voting systems instead of what would be most practical for this Donation Election. A lot of potential voters and donors may not be interested in using a super fancy optimal but technically involved voting method, and be the kind of small inconvenience that might turn people off the whole enterprise.
Now, before all you Seeing Like a State fans come at me saying how legibility is the devil's work I think I'm just going to disagree with you pre-emptively.[1] Sometimes there is a tradeoff between fidelity and legibility, and too much weighting on illegible technocracy can engender a lack of trust and have severe negative consequences.
1. ^
Actually it's interesting that Glen references Scott as on his side, I think there's actually some tension between their positions. But that's probably a topic for another post/discussion
3
Jason
Won't people be motivated to disapprove vote orgs in all cause areas but their preferred one? That would seemingly reduce approval voting to FPTP as between cause areas in effect.
4
JWS 🔸
Well, the top 3 charities will get chosen, so there's no benefit to you only selecting 1 option alone unless you really do believe only that 1 charity ought to get funded. I think AV may be more robust to these concerns than some think,[1] all I think all voting systems will have these edge cases.
I also may be willing simply bite the bullet here and trade-off a bit of strategic voting for legibility. But again, I don't think approval is worse than this than many other voting methods.
But my fundamental objection is that this is primarily a normative problem, where we want to be a community who'll vote honestly and not strategically. If GWWC endorse approval voting, then when you submit your votes there could be a pop-up with "I pledge not to vote strategically" or something like that.
I don't think any voting system is immune to that - Democracy works well because of the norms it spreads and trust it instills, as opposed to being the optimal transmission mechanism of individual preferences to a social welfare function imho.
1. ^
or here: https://link.springer.com/chapter/10.1007/978-1-4419-7539-3_2
5
Jason
Thanks. I assume there will be at least 3 orgs for each cause area.
If we can assume the forum is a "community who'll vote honestly and not strategically," approval voting would work -- but we shouldn't limit the winners to three in that case. Proportional representation among all orgs with net positive approval would be the fullest extent of the community's views, although some floor on support or cap on winners would be necessary for logistical reasons.
5
Jason
I'd prefer a voting mechanism that factored in as much of the vote as possible. I suspect that cause area will be a major determinant of individuals' votes, and would prefer that the voting structure promote engagement and participation for people with varying cause prioritizations.
Suppose we have 40% for cause A orgs, 25% for cause B orgs, 20% for cause C orgs, and 15% for various smaller causes. I would not prefer a method likely to select three organizations from cause A -- I don't think that outcome would be actually representative of the polis, and voting rules that would lead to such an outcome will discourage engagement and participation from people who sense that their preferred causes are not the leading one.
I'm not sure how to effectuate that preference in a voting system, although maybe people who have thought about voting systems more deeply than I could figure it out. I do think approval voting would be problematic; some voters might strategically disapprove all candidates except in their preferred cause area, which could turn the election into a cause-area election rather than an organization-specific one. Otherwise, it might be appropriate to assign each organization to a cause area, and provide that (e.g.) no more than half of all funds will go to organizations in the same cause area. If that rule were invoked, it would likely require selecting additional organizations than the initial three.
4
Jason
The more I think about this, the more I'd like at least one winner to be selected randomly among orgs that reach a certain vote threshold -- unsure if it should be weighted by vote total or equal between orgs. Maybe that org gets 15 to 20 percent of the take? That's a legible way to keep minority voices engaged despite knowing their preferences won't end up reflected in the top three.
4
Nathan Young
Proportional voting with some number of votes. between 1 and 10.
If it were me, the thing I'd experiment on is being able to donate votes to someone else. That feels like something I'd like to see more of on a larger scale. I give a vote to Jenifer and Alan, she researches longterm stuff, he looks into animal welfare.
5
abrahamrowe
FWIW, I mildly disagree with this, because a major part of the appeal of donation elections stuff (if done well) is that the results more closely model a community consensus than other giving mechanisms, and being able to donate votes would distort that in some sense. I think I don't see the appeal of being able to donate votes in this context over just telling Jenifer + Alan that they can control where one donates to some extent, or donating to a fund. Or, if not donating to the election fund, just asking Jenifer + Alan for their opinion and changing your own mind accordingly.
1
harfe
edit: I should have read the post more carefully
Do you intend to have one final winner or would it be ok to pay out the fund to various charities in different proportions (maybe with a minimum payout to avoid logistical hassle)? In the latter case, a consideration could also be proportional voting. But it is not clear how approval voting and ranked choice would work exactly in those cases.
Also, am I understanding correctly that donating more to that fund does not get you additional votes?
3
Lizka
We're planning on having 3 winners, and we'll allocate the funding proportionally across those three winners. So e.g. if we do approval voting, and candidate A gets 5 votes, B gets 2, C gets 20, and D gets 25, and we're distributing $100, then A (5 votes), C (20 votes), and D win (25 votes) and we'd send $10 to A, $40 to C, and $50 to D. I think this would straightforwardly work with quadratic voting (each person just has multiple vote-points). I haven't thought enough about how "proportional" allocation would work with ranked-choice votes.
And yep, donating more to that fund won't get you additional votes.
On October 27, 1962 (during the Cuban Missile Crisis), the Russian diesel-powered submarine B-59 started experiencing[1] nearby depth charges from US forces above them; the submarine had been detected and US ships seemed to be attacking. The submarine’s air conditioning was broken,[2] CO2 levels were rising, and B-59 was out of contact with Moscow. Two of the senior officers on the submarine, thinking that a global war had started, wanted to launch their “secret weapon,” a 10-kiloton nuclear torpedo. The captain, Valentin Savistky, apparently exclaimed: “We’re gonna blast them now! We will die, but we will sink them all — we will not become the shame of the fleet.”
The ship was authorized to launch the torpedo without confirmation from Moscow, but all three senior officers on the ship had to agree.[3] Chief of staff of the flotilla Vasili Arkhipov refused. He convinced Captain Savitsky that the depth charges were signals for the Soviet s... (read more)
For anyone interested in watching a dramatic reconstruction of this incident, go to timestamp 43:30–47:05 of The Man Who Saved The World. (I recommend watching at 1.5x speed.)
If you feel overwhelmed by FTX-collapse-related content on the Forum, you can hide most of it by using a tag filter: hover over the "FTX collapse" tag on the Frontpage (find it to the right of the "Frontpage Posts" header), and click on "Hidden."
[Note: this used to say "FTX crisis," and that might still show up in some places.]
Superman gets to business [private submission to the Creative Writing Contest from a little while back]
“I don’t understand,” she repeated. “I mean, you’re Superman.”
“Well yes,” said Clark. “That’s exactly why I need your help! I can’t spend my time researching how to prioritize while I should be off answering someone’s call for help.”
“But why prioritize? Can’t you just take the calls as they come?”
Lois clicked “Send” on the email she’d been typing up and rejoined the conversation. “See, we realized that we’ve been too reactive. We were taking calls as they came in without appreciating the enormous potential we had here. It’s amazing that we get to help people who are being attacked, help people who need our help, but we could also make the world safer more proactively, and end up helping even more people, even better, and when we realized that, when that clicked—”
“We couldn’t just ignore it.”
Tina looked back at Clark. “Ok, so what you’re saying is that you want to save people— or help people — and you think there are better and worse ways you could approach that, but you’re not sure which are which, and you realized that instead of rushing off to fight the most immediate threat, yo... (read more)
I recently ran a quick Fermi workshop, and have been asked for notes several times since. I've realized that it's not that hard for me to post them, and it might be relatively useful for someone.
Quick summary of the workshop
What is a Fermi estimate?
Walkthrough of the main steps for Fermi estimation
Notice a question
Break it down into simpler sub-questions to answer first
Don’t stress about the details when estimating answers to the sub-questions
Consider looking up some numbers
Put everything together
Sanity check
Different models: an example
Examples!
Discussion & takeaways
Resources
Guesstimate is a great website for Fermi estimation (although you can also use scratch paper or spreadsheets if that's what you prefer)
I attended and thoroughly enjoyed your workshop! Thanks for posting these notes
1
Lizka
Thanks for coming to the workshop, and for writing this note!
1
david_reinstein
I don’t see mention of quantifying the uncertainty in each component and aggregating this (usually via simulation). Is this not fundamental to Fermi? (Is it only a special version of Fermi, the “Monte Carlo” version?)
1
Lizka
Uncertainty is super important, and it's really useful to flag. It's possible I should have brought it up more during the workshop, and I'll consider doing that if I ever run something similar.
However, I do think part of the point of a Fermi estimate is to be easy and quick.
In practice, the way I'll sometimes incorporate uncertainty into my Fermis is by running the numbers in three ways:
1. my "best guess" for every component (2 hours of podcast episode, 100 episodes),
2. the "worst (reasonable) case" for every component (only 90? episodes have been produced, and they're only 1.5 hours long, on average), and
3. the "best case" for every component (150 episodes, average of 3 hours).
Then this still takes very little time and produces a reasonable range: ~135 to 450 hours of podcast (with a best guess of 200 hours). (Realistically, if I were taking enough care to run the numbers 3 times, I'd probably put more effort into the "best guess" numbers I produced.) I also sometimes do something similar with a spreadsheet/more careful Fermi.
I could do something more formal with confidence intervals and the like, and it's truly possible I should be doing that. But I really think there's a lot of value in just scratching something rough out on a sticky note during a conversation to e.g. see if a premise that's being entertained is worth the time, or to see if there are big obvious differences that are being missed because the natural components being considered are clunky and incompatible (before they're put together to produce the numbers we actually care about).
1
david_reinstein
Note that tools like Causal and Guesstimate make including uncertainty pretty easy and transparent.
I agree, but making uncertainty explicit makes it even better. (And I think it's an important epistemic/numeracy thing to cultivate and encourage). So I think if you are giving a workshop you should make this part of it at least to some extent.
1
david_reinstein
I think this would be worth digging into. It can make a big difference and it’s a mode we should be moving towards IMO, and should this be at the core of our teaching and learning materials. And there are ways of doing this that are not so challenging.
(Of course maybe in this particular podcast example it is now so important but in general I think it’s VERY important.)
“Worst case all parameters” is very unlikely. So is “best case everything”.
See the book “how to measure everything” for a discussion. Also the Causal and Guesstimate apps.
Moderation update: We have indefinitely banned 8 accounts[1] that were used by the same user (JamesS) to downvote some posts and comments from Nonlinear and upvote critical content about Nonlinear. Please remember that voting with multiple accounts on the same post or comment is very much against Forum norms.
(Please note that this is separate from the incident described here)
Was emerson_fartz an acceptable username in the first place? (It may not have had a post history in which case no one may have noticed its existence before the sockpuppeting detection, but that sounds uncivil toward a living person)
9
Lorenzo Buonanno🔸
It was not, and indeed it was only used for voting, so we noticed it only during this investigation
LukeDing (and their associated alt account) has been banned for six months, due to voting & multiple-account-use violations. We believe that they voted on the same comment/post with two accounts more than two hundred times. This includes several instances of using an alt account to vote on their own comments.
LukeDing appealed the decision; we will reach out to them and ask them if they’d like us to feature a response from them under this comment.
As some of you might realize, some people on the moderation team have conflicts of interest with LukeDing, so we wanted to clarify our process for resolving this incident. We uncovered the norm violation after an investigation into suspicious voting patterns, and only revealed the user’s identity to part of the team. The moderators who made decisions about how to proceed weren’t aware of LukeDing's identity (they only saw anonymized information).
Is more information about the appellate process available? The guide to forum norms says "We're working on a formal process for reviewing submissions to this form, to make sure that someone outside of the moderation team will review every submission, and we'll update this page when we have a process in place."
The basic questions for me would include: information about who decides appeals, how much deference (if any) the adjudicator will give to the moderators' initial decision -- which probably should vary based on the type of decision at hand, and what kind of contact between the mods and appellate adjudicator(s) is allowed. On the last point, I would prefer as little ex parte contact if possible, and would favor having an independent vetted "advocate for the appellant" looped in if there needs to be contact to which the appellant is not privy.
Admittedly I have a professional bias toward liking process, but I would err on the side of more process than less where accounts are often linked to real-world identities and suspensions are sometimes for conduct that could be seen as dishonest or untrustworthy. I would prefer public disclosure of an action taken in cases like this only after the appellate process is complete for the same reasons, assuming the user timely indicates a desire to appeal the finding of a norm violation.
Finally, I commend keeping the moderators deciding whether a violation occurred blinded as to the user's identity as a best practice in cases like this, even where there are no COIs. It probably should be revealed prior to determining a sanction, though.
7
Larks
It does intuitively seem like an immediate temporary ban, made public only after whatever appeals are allowed have been exhausted, should give the moderation team basically everything they need while being more considerate of anyone whose appeals are ultimately upheld (i.e. innocent, or mitigating circumstances).
Moderation update: A new user, Bernd Clemens Huber, recently posted a first post ("All or Nothing: Ethics on Cosmic Scale, Outer Space Treaty, Directed Panspermia, Forwards-Contamination, Technology Assessment, Planetary Protection, (and Fermi's Paradox)") that was a bit hard to make sense of. We hadn't approved the post over the weekend and hadn't processed it yet, when the Forum team got an angry and aggressive email today from the user in question calling the team "dipshits" (and providing a definition of the word) for waiting throughout the weekend.
If the user disagrees with our characterization of the email, they can email us to give permission for us to share the whole thing.
We have decided that this is not a promising start to the user's interactions on the Forum, and have banned them indefinitely. Please let us know if you have concerns, and as a reminder, here are the Forum's norms.
We have strong reason to believe that Torres (philosophytorres) used a second account to violate their earlier ban. We feel that this means that we cannot trust Torres to follow this forum’s norms, and are banning them for the next 20 years (until 1 October 2042).
Moderation update: We have banned "Richard TK" for 6 months for using a duplicate account to double-vote on the same posts and comments. We’re also banning another account (Anin, now deactivated), which seems to have been used by that same user or by others to amplify those same votes. Please remember that voting with multiple accounts on the same post or comment is very much against Forum norms.
(Please note that this is separate from the incident described here)
We’re issuing [Edit: identifying information redacted] a two-month ban for using multiple accounts to vote on the same posts and comments, and in one instance for commenting in a thread pretending to be two different users. [Edit: the user had a total of 13 double-votes, most far apart and are likely accidental, two upvotes close together on others' posts (which they claim are accidental as well), but two cases of deliberate self upvote from alternative accounts]
This is against the Forumnorms around using multiple accounts. Votes are really important for the Forum: they provide feedback to authors and signal to readers what other users found most valuable, so we need to be particularly strict in discouraging this kind of vote manipulation.
A note on timing: the comment mentioned above is 7 months old but went unnoticed at the time, a report for it came in last week and triggered this investigation.
If [Edit: redacted] thinks that this is not right, he can appeal. As a reminder, bans affect the user, not the account.
[Edit: We have retroactively decided to redact the user's name from this early message, and are currently rethinking our policies on the mat... (read more)
[A moderator had edited this comment to remove identifying information, after a moderation decision to retroactively redact the user's identification]
7
Linch
I guess it makes sense that people who disagree with the norms are more likely to do underhanded things to violate them.
4
Lorenzo Buonanno🔸
Just quickly noting that none of the double-votes were on that thread or similar ones, as far as I know.
2
Jason
Do suspended users get a chance to make a public reply to the mod team's findings? I don't think that's always necessary -- e.g., we all see the underlying conduct when public incivility happens -- but I think it's usually warranted when the findings imply underhanded behavior ("pretending") and the underlying facts aren't publicly observable. There's an appeal process, but that doesn't address the public-reputation interests of the suspended person.
We have strong reason to believe that Charles He used multiple new accounts to violate his earlier 6-month-long ban. We feel that this means that we cannot trust Charles He to follow this forum’s norms, and are banning him from the Forum for the next 10 years (until December 20, 2032).
We have already issued temporary suspensions to several suspected duplicate accounts, including one which violated norms about rudeness and was flagged to us by multiple users. We will be extending the bans for each of these accounts to mirror Charles’s 10-year ban, but are giving the users an opportunity to message us if we have made any of those temporary suspensions in error (and have already reached out to them). While we aren’t >99% certain about any single account, we’re around 99% that at least one of these is Charles He.
I find this reflects worse on the mod team than Charles. This is nowhere near the first time I've felt this way.
Fundamentally, it seems the mod team heavily prioritizes civility and following shallow norms above enabling important discourse. The post on forum norms says a picture of geese all flying in formation and in one direction is the desirable state of the forum; I disagree that this is desirable. Healthy conflict is necessary to sustain a healthy community. Conflict sometimes entails rudeness. Some rudeness here and there is not a big deal and does not need to be stamped out entirely. This also applies to the people who get banned for criticizing EA rudely, even when they're criticizing EA for its role in one of the great frauds of modern history. Banning EA critics for minor reasons is a short-sighted move at best.
Banning Charles for 10 years (!!) for the relatively small crime of evading a previous ban is a seriously flawed idea. Some of his past actions like doxxing someone (without any malice I believe) are problematic and need to be addressed, but do not deserve a 10 year ban. Some of his past comments, especially farther in the past, have been frustrating and net-negative to me, but these negative actions are not unrelated to some of his positive traits, like his willingness to step out of EA norms and communicate clearly rather than like an EA bot. The variance of his comments has steadily decreased over time. Some of his comments are even moderator-like, such as when he warned EA forum users not to downvote a WSJ journalist who wasn't breaking any rules. I note that the mod team did not step in there to encourage forum norms.
I also find it very troubling that the mod team has consistent and strong biases in how it enforces its norms and rules, such as not taking any meaningful action against an EA in-group member for repeated and harmful violations of norms but banning an EA critic for 20 years for probably relatively minor and harmless viola
the relatively small crime of evading a previous ban
I don't think repeatedly evading moderator bans is a "relatively small crime". If Forum moderation is to mean anything at all, it has to be consistently enforced, and if someone just decides that moderation doesn't apply to them, they shouldn't be allowed to post or comment on the Forum.
Charles only got to his 6 month ban via a series of escalating minor bans, most of which I agreed with. I think he got a lot of slack in his behaviour because he sometimes provided significant value, but sometimes (with insufficient infrequency) behaved in ways that were seriously out of kilter with the goal of a healthy Forum.
I personally think the 10-year thing is kind of silly and he should just have been banned indefinitely at this point, then maybe have the ban reviewed in a little while. But it's clear he's been systematically violating Forum policies in a way that requires serious action.
The post on forum norms says a picture of geese all flying in formation and in one direction is the desirable state of the forum; I disagree that this is desirable.
Indefinite suspension with leave to seek reinstatement after a stated suitable period would have been far preferable to a 10-year ban. A tenner isn't necessary to vindicate the moderators' authority, and the relevant conduct doesn't give the impression of someone for whom the passage of ten years' time is necessary before there is a reasonable probability that would they have become a suitable participant during the suspension.
2[anonymous]
It makes a lot of difference to me that Charles' behavior was consistently getting better. If someone consistently flouts norms without any improvement, at some point they should be indefinitely banned. This is not the case with Charles. He started off with really high variance and at this point has reached a pretty tolerable amount. He has clearly worked on his actions. The comments he posted while flouting the mods' authority generally contributed to the conversation. There are other people who have done worse things without action from the mod team. Giving him a 10 year ban without appeal for this feels more motivated by another instance of the mod team asserting their authority and deciding not to deal with messiness someone is causing than a principled decision.
5
Will Bradshaw
I think this is probably true. I still think that systematically evading a Forum ban is worse behaviour (by which I mean, more lengthy-ban-worthy) than any of his previous transgressions.
I am not personally aware of any, and am sceptical of this claim. Open to being convinced, though.
7
pseudonym
can you give some examples of this?
3[anonymous]
Various comments made by this user in multiple posts some time ago, some of which received warnings by mods but nothing beyond that.
2
Ramiro
Totally unrelated to the core of the matter, but do you intend to turn this into a frontpage post? I'm a bit inclined to say it'd be better for transparency, and to inform others about the bans, and deter potential violators.... but I'm not sure, maybe you have a reason for preferring the shortform (or you'll publish periodical updates on the frontpage
It's kind of jarring to read that someone has been banned for "violating a norm" - that word to me implies that they're informal agreements between the community. Why not call them "rules"?
pinkfrog (and their associated account) has been banned for 1 month, because they voted multiple times on the same content (with two accounts), including upvoting pinkfrog's comments with their other account. To be a bit more specific, this happened on one day, and there were 12 cases of double-voting in total (which we’ll remove). This is against our Forum norms on voting and using multiple accounts.
As a reminder, bans affect the user, not the account(s).
If anyone has questions or concerns, please feel free to reach out, and if you think we made a mistake here, you can appeal the decision.
Multiple people on the moderation team have conflicts of interest with pinkfrog, so I wanted to clarify our process for resolving this incident. We uncovered the norm violation after an investigation into suspicious voting patterns, and only revealed the user’s identity to part of the team. The moderators who made decisions about how to proceed aren't aware of pinkfrog's real identity (they only saw anonymized information).
It seems inconsistent to have this info public for some, and redacted for others. I do think it is good public service to have this information public, but am primarily pushing here for consistency and some more visibility around existing decisions.
Agree. It seems potentially pretty damaging to people’s reputations to make this information public (and attached to their names); that strikes me as a much bigger penalty than the bans. There should, at a minimum, be a consistent standard, and I’m inclined to think that standard should be having a high bar for releasing identifying information.
I think we should hesitate to protect people from reputational damage caused by people posting true information about them. Perhaps there's a case to be made when the information is cherry-picked or biased, or there's no opportunity to hear a fair response. But goodness, if we've learned anything from the last 18 months I hope it would include that sharing information about bad behaviour is sometimes a public good.
I would guess that most people engage in private behavior that would be reputationally damaging if the internet were to find out about it. Just because something is true doesn’t mean you forfeit your rights to not have that information be made public.
I think people might reasonably (though wrongly) assume that forum mods are not monitoring accounts at this level of granularity, and thus believe that their voting behavior is private. Given this, I think mods should warn before publicly censoring. (Just as it would be better to inform your neighbor that you can see them doing something embarrassing through their window before calling the police or warning other people about then—maybe they just don’t realize you can see, and telling them is all they need to not do the thing anymore, which, after all, is the goal.)
Frankly, I don’t love that mods are monitoring accounts at this level of granularity. (For instance, knowing this would make me less inclined to put remotely sensitive info in a forum dm.)
6
Will Aldred
Writing in a personal capacity; I haven't run this by other mods.
Hi, just responding to these parts of your comment:
We include some detail on what would lead moderators to look into a user's voting activity, and what information we have access to, on our "Guide to norms on the Forum" page:
(In addition, note that moderators can't just go into a user's account and check their voting history even when we do have reason to look into that user. We require one of the Forum engineers to run some queries on the back end to yield this information.)
Finally, to address your concern about direct messages on the Forum: like a regular user, a moderator cannot see into anyone else's messages.
Hope this is helpful :)
8
Lorenzo Buonanno🔸
Also writing in a personal capacity.
Thanks for writing this! To clarify a few points even more:
I confirm this, and just want to highlight that
1. this is pretty rare; we have a high bar before asking developers to look into patterns
2. usually, one developer looks into things, and shares anonymized data with moderators, who then decide whether it needs to be investigated more deeply
3. If so, a subset of moderators gets access to deanonymized data to make a decision and contact/warn/ban the user(s)
On
I confirm this, but I want to highlight that messages on the forum are not end-to-end encrypted and are, by default, sent via email as well (i.e. when you get a message on the forum you also get an email with the message). So forum developers and people who have or will have access to the recipient's email inbox, or the forum's email delivery service, can see the messages.
For very private communications, I would recommend using privacy-first end-to-end encrypted platforms like Signal.
6
lilly
Thanks; this is helpful and reassuring, especially re: the DMs. I had read this section of the norms page, and it struck me that the “if we have reason to believe that someone is violating norms around voting” clause was doing a lot of work. I would appreciate more clarification about what would lead mods to believe something like this (and maybe some examples of how you’ve come to have such beliefs). But this is not urgent, and thanks for the clarification you’ve already provided.
8
Will Aldred
Yeah, this is a reasonable thing to ask. So, the “if we have reason to believe that someone is violating norms around voting” clause is intentionally vague, I believe, because if we gave more detail on the kinds of checks/algorithms we have in place for flagging potential violations, then this could help would-be miscreants commit violations that slip past our checks.
(I’m a bit sad that the framing here is adversarial, and that we can’t give users like you more clarification, but I think this state of play is the reality of running an online forum.)
If it helps, though, the bar for looking into a user’s voting history is high. Like, on average I don’t think we do this more than once or twice per month.
4
lilly
Thanks, this is also helpful! One thing to think about (and no need to tell me), is whether making the checks public could effectively disincentivize the bad behavior (like how warnings about speed cameras may as effectively disincentivize speeding as the cameras do themselves). But if there are easy workarounds, I can see why this wouldn’t be viable.
2
Ben Millwood🔸
I agree that not all true things should be made public, but I think when it specifically pertains to wrongdoing and someone's trustworthiness, the public interest can override the right to privacy. If you look into your neighbour's window and you see them printing counterfeit currency, you go to the police first, rather than giving them an opportunity to simply hide their fraud better.
8
lilly
Maybe the crux is: I think forum users upvoting their own comments is more akin to them Facetuning dating app photos than printing counterfeit currency. Like, this is pretty innocuous behavior and if you just tell people not to do it, they’ll stop.
2
Ben Millwood🔸
It seems like we disagree on how bad it is to self-vote (I don't think it's anywhere near the level of "actual crime", but I do think it's pretty clearly dishonest and unfair, and for such a petty benefit it's hard for me to feel sympathetic to the temptation).
But I don't think it's the central point for me. If you're simultaneously holding that:
* this information isn't actually a big deal, but
* releasing this publically would cause a lot of harm through reputational damage,
then there's a paternalistic subtext where people can't be trusted to come to the "right" conclusions from the facts. If this stuff really wasn't a big deal, then talking about it publically wouldn't be a big deal either. I don't think people should be shunned forever and excluded from any future employment because they misused multiple accounts on the forum. I do think they should be a little embarrassed, and I don't think that moving to protect them from that embarrassment is actually a kindness from a community-wide perspective.
4
lilly
I feel like this is getting really complicated and ultimately my point is very simple: prevent harmful behavior via the least harmful means. If you can get people to not vote for themselves by telling them not to, then just… do that. I have a really hard time imagining that someone who was warned about this would continue to do it; if they did, it would be reasonable to escalate. But if they’re warned and then change their behavior, why do I need to know this happened? I just don’t buy that it reflects some fundamental lack of integrity that we all need to know about (or something like this).
5
Ben Millwood🔸
I think that posting that someone is banned and why they were banned is not mainly about punishing them. It's about helping people understand what the moderation team is doing, how rule-breaking is handled, and why someone no longer has access to the forum. For example, it helps us to understand if the moderation team are acting on inadequate information, or inconsistently between different people. The fact that publishing this information harms people is an unfortunate side effect, after the main effect of improving transparency and keeping people informed.
It doesn't even really feel right to call them harmed by the publication. If people are harmed by other people knowing they misuse the voting system, I'd say they were mainly harmed by their own misuse of the system, not by someone reporting on it.
Then you needn't object to the moderation team talking about what they did!
6
Jason
It's unclear to me that naming names materially advances the first two goals. As to the third, the suspended user could have the option of having their name disclosed. Otherwise, I don't think we're entitled to an explanation of why a particular poster isn't active anymore.
4
Jason
There's also the interest in deterring everyone else from doing it (general deterrence), not just in getting these specific people to stop doing it (specific deterrence). While I have mixed feelings about publicly naming offenders, the penalty does need to sting enough to make the benefits of the offense not worth the risk of getting caught. A private warning with no real consequences might persuade the person violating the rules not to do it again, but double-voting would surge as people learned you get a freebie.
2
lilly
“double-voting would surge as people learned you get a freebie.”
I just don’t see this happening?
Separately, one objection I have to cracking down hard on self-voting is that I think this is not very harmful relative to other ways in which people don’t vote how they’re “supposed to.” E.g., we know the correlation between upvotes and agree votes is incredibly high, and downvoting something solely because you disagree with it strikes me as more harmful to discourse on the forum than self-voting. I think the reason self-voting gets highlighted isn’t because it’s especially harmful, it’s just because it’s especially catchable.
If the mods want to improve people’s voting behavior on the forum, I both wish they’d target different voting behavior (ie, the agree/upvoting correlation) and use different means to do it (ie, generating reports for people of their own voting correlations, whether they tend to upvote/downvote certain people, etc), rather than naming/shaming people for self-voting.
4
Jason
I think it's more that upvoting your own posts from an alt is (1) willful, intentional behavior (2) aimed at deceiving the community about the level of support of a comment (3) for the person's own benefit. Presumably, most people who are doing it are employing some sort of means to evade detection, which adds another layer of deceptiveness. While I don't like downvoting-for-disagreement and the like either, that kind of behavior presumptively reflects a natural cognitive bias rather than any of the three characteristics listed above. It is for those reasons that -- in my view -- downvoting-for-disagreement is generally not the proper subject of a sanctioning system,[1] while self-upvoting is.
I've suggested to the mods before that sanctions should sometimes be more carefully tailored to the offense, so I'd be open to the view that consequences like permanently denying the violator's ability to vote and their ability to use alts might be more tailored to the offense than public disclosure. Those are the specific functions which they have demonstrated an inability to handle responsibly. Neither function is so fundamental to the ability to use the Forum that the mods should feel obliged to expend their time deciding if the violator has rehabilitated themselves enough to restore those privileges.
1. ^
There could be circumstances in which soft-norm violative behavior was so extreme that sanctions should be considered. However, unlike "don't multi-vote" (which is a bright-line rule for which the violator should be perfectly aware that they are violating the rules), these norms are less clearcut -- so privately reaching out to the person would be the appropriate first action in a case like that.
6
pseudonym
Fair point about reputational harms being worse and possibly too punishing in some cases. I think in terms of a proposed standard it might be worth differentiating (if possible) between e.g. careless errors, or momentary lapses in judgement that were quickly rectified and likely caused no harm in expectation, versus a pattern of dishonest voting intended to mislead the EAF audience, and especially if they or an org that they work for stand to gain from it, or the comments in question are directly harmful to another org. In these latter cases the reputational harm may be more justifiable.
8
Jason
For reasoning transparency / precedent development, it might be worthwhile to address two points:
(1) I seem to remember other multivoting suspensions being much longer than 1 month. I had gotten the impression that the de facto starting point for deliberate multiaccount vote manipulation was ~ six months. Was the length here based on mitigating factors, perhaps the relatively low number of violations and that they occurred on a single day? If the usual sanction is ~ six months, I think it would be good to say that here so newer users understand that multivoting is a really big deal.
(2) Here the public notice names the anon account pinkfrog (which has 3 comments + 50 karma), rather than the user's non-anon account. The last multi account voting suspension I saw named the user's primary account, which was their real name. Even though the suspension follows the user, which account is publicly named can have a significant effect on public reputation. How does the mod team decide which user to name in the public notice?
6
pseudonym
pinkfrog: 1 month (12 cases of double voting)
LukeDing: 6 months (>200 times)
JamesS: indefinite (8 accounts, number not specified)
[Redacted]: 2 months (13 double votes, most are "likely accidental", two "self upvotes")
RichardTK: 6 months (number not specified)
Charles He: 10 years (not quite analogous as these are using alts to circumvent initial bans, included other violations)
Torres: 20 years (not quite analogous as these are using alts to circumvent initial bans, included other violations)
2[anonymous]
Torres was banned for 20 years according to the link.
2
pseudonym
Corrected, thanks!
6
Lorenzo Buonanno🔸
Reply to this comment from @John G. Halstead
(Written in a personal capacity, I did not check this with other moderators)
Thank you for the feedback! I didn't want to go too off-topic, as this is unrelated to this post, so I'm replying here, but I want to quickly share some factual information for other readers.
You're writing this in multiple comments. I want to make it clear that moderators did not endorse or "defend" (or symmetrically "attack") the post as moderators. But of course, we do comment as users on parts we agree or disagree with (like any other user). Let us know if it's not clear whether we're commenting as users or as moderators.
As for your other warnings, I want to make sure other readers know that your last warning was not for discussing a specific topic, but for being uncivil and not constructive to the discussion. I agree that the situation in the first warning is less relevant to this case, apologies for bringing it up.
6
Lizka
Just a quick note to say that we’ve removed a post sharing a Fermi estimate of the chances that the author finds a partner who matches their preferred characteristics and links to a date-me doc.
The Forum is for discussions about improving the world, and a key norm we highlight is “Stay on topic.” This is not the right space for coordinating dating. (Consider exploring LessWrong, ACX threads/classifieds, or EA-adjacent Facebook/Reddit/Discord groups for discussions that are primarily social.)
We’re not taking any other action about the author, although I've asked them to stay on topic in the future.
4
Lizka
Quick update: we’ve banned Defacto, who we have strong reason to believe is another sockpuppet account for Charles He. We are extending Charles’s ban to be indefinite (he and others can appeal if they want to).
You can find more on our rules for pseudonymity and multiple accounts here. If you have any questions or concerns about this, please also feel free to reach out to us at forum-moderation@effectivealtruism.org.
6
Lizka
Moderation update:
Around a month ago, a post about the authorship of Democratising Risk got published. This post got taken down by its author. Before this happened, the moderation team had been deciding what to do with some aspects of the post (and the resulting discussion) that had violated Forum norms. We were pretty confident that we’d end up banning two users for at least a month, so we banned them temporarily while we sorted some things out.
One of these users was Throwaway151. We banned them for posting something a bit misleading (the post seemed to overstate its conclusions based on the little evidence it had, and wasn’t updated very quickly based on clear counter-evidence), and being uncivil in the comments. Their ban has passed, now. As a reminder, bans affect the user, not the account, so any other accounts Throwaway151 operated were also affected. The other user was philosophytorres — see the relevant update.
4
Lizka
We’ve banned Vee from the Forum for 1 year. Their content seems to be primarily or significantly AI-generated,[1] and it’s not clear that they’re using it to share thoughts they endorse and have carefully engaged with. (This had come up before on one of their posts.) Our current policy on AI-generated content makes it clear that we’ll be stricter when moderating AI-generated content. Vee’s content doesn’t meet the standards of the Forum.
If Vee thinks that this is not right, they can appeal. If they come back, we’ll be checking to make sure that their content follows Forum norms. As a reminder, bans affect the user, not the account.
1. ^
Different detectors for AI content are giving this content different scores, but we think that this is sufficiently likely true to act on.
It’s hard to be certain that something is AI-generated, and I’m not very satisfied with our processes or policies on this front. At the same time, the increase in the number of bots has made dealing with spam or off-topic/troll contributions harder, and I think that waiting for something closer to certainty will have costs that are too high.
6
JP Addison🔸
Update, we have unbanned Vee. We are new to using AI detection tools and we made a mistake. We apologize.
4
Lorenzo Buonanno🔸
Moderation update:
I'm indefinitely banning JasMaguire for an extremely racist comment that has since been deleted. We'll likely revisit and update our forum norms to explicitly discourage this sort of behavior.
Please feel free to get in touch with forum-moderation@effectivealtruism.org if you have any concerns.
-1
JP Addison🔸
Moderation update: We're issuing dstudiocode a one-month ban for breaking Forum norms in their recent post and subsequent behavior. Specifically:
* Posting content that could be interpreted as promoting violence or illegal activities.
The post in question, which asked whether murdering meat-eaters could be considered "ethical," crosses a line in terms of promoting potential violence.
As a reminder, the ban affects the user, not just the account. During their ban period, the user will not be permitted to rejoin the Forum under another account name. If they return to the Forum after the ban period, we'll expect a higher standard of norm-following and compliance with moderator instructions.
You can reach out to forum-moderation@effectivealtruism.org with any questions. You can appeal this decision here.
I suggest editing this comment to note the partial reversal on appeal and/or retracting the comment, to avoid the risk of people seeing only it and reading it as vaguely precedential.
Strong +1 to Richard, this seems a clear incorrect moderation call and I encourage you to reverse it.
I'm personally very strongly opposed to killing people because they eat meat, and the general ethos behind that. I don't feel in the slightest offended or bothered by that post, it's just one in a string of hypothetical questions, and it clearly is not intended as a call to action or to encourage action.
If the EA Forum isn't somewhere where you can ask a perfectly legitimate hypothetical question like that, what are we even doing here?
The moderators have reviewed the decision to ban @dstudioscode after users appealed the decision. Tl;dr: We are revoking the ban, and are instead rate-limiting dstudioscode and warning them to avoid posting content that could be perceived as advocating for major harm or illegal activities. The rate limit is due to dstudiocode’s pattern of engagement on the Forum, not simply because of their most recent post—for more on this, see the “third consideration” listed below.
More details:
Three moderators,[1] none of whom was involved in the original decision to ban dstudiocode, discussed this case.
The first consideration was “Does the cited norm make sense?” For reference, the norm cited in the original ban decision was “Materials advocating major harm or illegal activities, or materials that may be easily perceived as such” (under “What we discourage (and may delete or edit out)” in our “Guide to norms on the Forum”). The panel of three unanimously agreed that having some kind of Forum norm in this vein makes sense.
The second consideration was “Does the post that triggered the ban actually break the cited norm?” For reference, the post ended with the question “sho... (read more)
I appreciate the thought that went into this. I also think that using rate-limits as a tool, instead of bans, is in general a good idea. I continue to strongly disagree with the decisions on a few points:
I still think including the "materials that may be easily perceived as such" clause has a chilling effect.
I also remember someone's comment that the things you're calling "norms" are actually rules, and it's a little disingenuous to not call them that; I continue to agree with this.
The fact that you're not even willing to quote the parts of the post that were objectionable feels like an indication of a mindset that I really disagree with. It's like... treating words as inherently dangerous? Not thinking at all about the use-mention distinction? I mean, here's a quote from the Hamas charter: "There is no solution for the Palestinian question except through Jihad." Clearly this is way way more of an incitement to violence than any quote of dstudiocode's, which you're apparently not willing to quote. (I am deliberately not expressing any opinion about whether the Hamas quote is correct; I'm just quoting them.) What's the difference?
"They see the fact that it is “just” a philosophical
I'm pretty sure we could come up with various individuals and groups of people that some users of this forum would prefer not to exist. There's no clear and unbiased way to decide which of those individuals and groups could be the target of "philosophical questions" about the desirability of murdering them and which could not. Unless we're going to allow the question as applied to any individual or group (which I think is untenable for numerous reasons), the line has to be drawn somewhere. Would it be ethical to get rid of this meddlesome priest should be suspendable or worse (except that the meddlesome priest in question has been dead for over eight hundred years).
And I think drawing the line at we're not going to allow hypotheticals about murdering discernable people[1] is better (and poses less risk of viewpoint suppression) than expecting the mods to somehow devise a rule for when that content will be allowed and consistently apply it. I think the effect of a bright-line no-murder-talk rule on expression of ideas is modest because (1) posters can get much of the same result by posing non-violent scenarios (e.g., leaving someone to drown in a pond is neither an a... (read more)
Do you think it is acceptable to discuss the death penalty on the forum? Intuitively this seems within scope - historically we have discussed criminal justice reform on the forum, and capital punishment is definitely part of that.
If so, is the distinction state violence vs individual violence? This seems not totally implausible to me, though it does suggest that the offending poster could simply re-word their post to be about state-sanctioned executions and leave the rest of the content untouched.
7
NickLaing
I've weak karma downvoted and disagreed with this, then hit the "insightful" button. Definitely made me think and learn.
I agree that this is really tricky question, and some of those philosophical conversations (including this one) are important and should happen, but I don't think this particular EA forum is the best place for them, for a few reasons.
1) I think there are better places to have these often awkward, fraught conversations. I think they are often better had in-person where you can connect, preface, soften and easily retract. I recently got into a mini online-tiff, when a wise onlooker noted...
"Online discussions can turn that way with a few misinterpretations creating a doom loop that wouldn't happen with a handshake and a drink"
Or alternatively perhaps in a more academic/narrow forum where people have similar discussion norms and understandings. This forum has a particularly wide range of users, from nerds to philosophers to practitioners to managers to donors so there's a very wide range of norms and understandings.
2) There's potential reputational damage for all the people doing great EA work across the spectrum here. These kinds of discussions could lead to more hit-pieces and reduced funding. It would be a pity if the AI apocalypse hit us because of funding cuts due to these discussions. (OK now I'm strawmanning a bit :D)
3) The forum might be an entry-point for some people into EA things. I don't think its a good idea for this to be these discussions to be the first thing someone looking into EA sees on the internet.
4) It might be a bit of a strawman to say our "discourse will forever be at the mercy of whoever is most hostile to you, or whoever is craziest." I think people hostile to EA don't like many things said here on the forum, but we aren't forever at the mercy of them and we keep talking. I think there are just a few particular topics which give people more ammunition for public take-downs, and there is wisdom in sometimes avo
I think there are better places to have these often awkward, fraught conversations.
You are literally talking about the sort of conversations that created EA. If people don't have these conversations on the forum (the single best way to create common knowledge in the EA commmunity), then it will be much harder to course-correct places where fundamental ideas are mistaken. I think your comment proceeds from the implicit assumption that we're broadly right about stuff, and mostly just need to keep our heads down and do the work. I personally think that a version of EA that doesn't have the ability to course-correct in big ways would be net negative for the world. In general it is not possible to e.g. identify ongoing moral catastrophes when you're optimizing your main venue of conversations for avoiding seeming weird.
I agree with you the quote from the Hamas charter is more dangerous - and think we shouldn't be publishing or discussing that on the forum either.
If you're not able to talk about evil people and their ideologies, then you will not be able to account for them in reasoning about how to steer the world. I think EA is already far ... (read more)
Just to narrow in on a single point—I have found the 'EA fundamentally depends on uncomfortable conversations' point to be a bit unnuanced in the past. It seems like we could be more productive by delineating which kinds of discomfort we want to defend—for example, most people here don't want to have uncomfortable conversations about age of consent laws (thankfully), but do want to have them about factory farming.
When I think about the founding myths of EA, most of them seem to revolve around the discomfort of applying utilitarianism in practice, or on how far we should expand our moral circles. I think EA would've broadly survived intact by lightly moderating other kinds of discomfort (or it may have even expanded).
I'm not keen to take a stance on whether this post should or shouldn't be allowed on the forum, but I am curious to hear if and where you would draw this line :)
Narrowing in even further on the example you gave, as an illustration: I just had an uncomfortable conversation about age of consent laws literally yesterday with an old friend of mine. Specifically, my friend was advocating that the most important driver of crime is poverty, and I was arguing that it's cultural acceptance of crime. I pointed to age of consent laws varying widely across different countries as evidence that there are some cultures which accept behavior that most westerners think of as deeply immoral (and indeed criminal).
Picturing some responses you might give to this:
That's not the sort of uncomfortable claim you're worried about
But many possible continuations of this conversation would in fact have gotten into more controversial territory. E.g. maybe a cultural relativist would defend those other countries having lower age of consent laws. I find cultural relativism kinda crazy (for this and related reasons) but it's a pretty mainstream position.
I could have made the point in more sensitive ways
Maybe? But the whole point of the conversation was about ways in which some cultures are better than others. This is inherently going to be a sensitive claim, and it's hard
Do you have an example of the kind of early EA conversation which you think was really important which helped came up with core EA tenets might be frowned upon or censored on the forum now? I'm still super dubious about whether leaving out a small number of specific topics really leaves much value on the table.
And I really think conversations can be had in more sensitive ways. In the the case of the original banned post, just as good a philosophical conversation could be had without explicitly talking about killing people. The conversation already was being had on another thread "the meat eater problem"
And as a sidebar yeah I wouldn't have any issue with that above conversation myself because we just have to practically discuss that with donors and internally when providing health care and getting confronted with tricky situations. Also (again sidebar) it's interesting that age of marriage/consent conversations can be where classic left wing cultural relativism and gender safeguarding collide and don't know which way to swing. We've had to ask that question practically in our health centers, to decide who to give family planning to and when to think of referring to police etc. Super tricky.
My point is not that the current EA forum would censor topics that were actually important early EA conversations, because EAs have now been selected for being willing to discuss those topics. My point is that the current forum might censor topics that would be important course-corrections, just as if the rest of society had been moderating early EA conversations, those conversations might have lost important contributions like impartiality between species (controversial: you're saying human lives don't matter very much!), the ineffectiveness of development aid (controversial: you're attacking powerful organizations!), transhumanism (controversial, according to the people who say it's basically eugenics), etc.
Re "conversations can be had in more sensitive ways", I mostly disagree, because of the considerations laid out here: the people who are good at discussing topics sensitively are mostly not the ones who are good at coming up with important novel ideas.
For example, it seems plausible that genetic engineering for human intelligence enhancement is an important and highly neglected intervention. But you had to be pretty disagreeable to bring it into the public conversation a few years ago (I think it's now a bit more mainstream).
This moderation policy seems absurd. The post in question was clearly asking purely hypothetical questions, and wasn't even advocating for any particular answer to the question. May as well ban users for asking whether it's moral to push a man off a bridge to stop a trolley, or ban Peter Singer for his thought experiments about infanticide.
Perhaps dstudiocode has misbehaved in other ways, but this announcement focuses on something that should be clearly within the bounds of acceptable discourse. (In particular, the standard of "content that could be interpreted as X" is a very censorious one, since you now need to cater to a wide range of possible interpretations.)
Ah, thanks, that's important context - I semi-retract my strongly worded comment above, depending on exactly how bad the removed post was, but can imagine posts in this genre that I think are genuinely bad
4
JP Addison🔸
Another comment from me:
I don’t like my mod message, and I apologize for it. I was rushed and used some templated language that I knew damn well at the time that I wasn’t excited about putting my name behind. I nevertheless did and bear the responsibility.
That’s all from me for now. The mods who weren’t involved in the original decision will come in and reconsider the ban, pursuant to the appeal.
-5
JP Addison🔸
9
Jason
I'd separate out the removal and the suspension, and dissent only as to the latter.
I get why the mods would feel the need to chart a wide berth around anything that some person could somehow "interpret[] as promoting violence or illegal activities." Making a rule against brief hypothetical mentions of the possible ethics of murder is defensible, especially in light of certain practical realities.
However, I can't agree with taking punitive action against a user where the case that they violated the norm is this tenuous and there is a lack of fair prior notice of the mods' interpretation. For that kind of action, I think the minimum standard would be either clear notice or content that a reasonable person would recognize could reasonably be interpreted as promoting violence. In other words, was the poster negligent in failing to recognize that violence promotion was a reasonable interpretation?
I don't think the violence-promoting interpretation is a reasonable one here, and it sounds like several other users agree -- which I take as evidence of non-negligence.
3-4 minutes, mostly on playing through various elimination-order scenarios in my head and trying to ensure that my assigned values would still reflect my preferences in at least more likely scenarios.
4
Kaleem
took me ~5min
4
Will Aldred
It took me just under 5 minutes.
The percentages I inputted were best guesses based on my qualitative impressions. If I'd been more quantitative about it, then I expect my allocations would have been better—i.e., closer to what I'd endorse on reflection. But I didn't want to spend long on this, and figured that adding imperfect info to the commons would be better than adding no info.
4
Daniel_Eth
IIRC it took me about a minute or two. But I already had high context and knew how I wanted to vote, so after getting oriented I didn't have to spend time learning more or thinking through tradeoffs.
3
Will Howard🔹
It took me ~1 minute. I already had a favourite candidate so I put all my points towards that. I was half planning to come back and edit to add backup choices but I've seen the interim results now so I'm not going to do that.
2
Nathan Young
Probably about 30 minutes of unfocused thought on the actual voting. Mainly it was spent negotiating between what I thought was sort of best and some guilt and status based obligation stuff.
On top of that I perhaps read 2-4 articles and chatted to 1-2 people involved in orgs. I guess that was 1- 3 hours.
1
Toby Tremlett🔹
I think around 5-10 mins? I tried to compare everything I cared at all about, so I only used multipliers between 0 and 2 (otherwise I would have lost track and ended up with intransitive preferences). The comparison stage took the most time. I edited things in the end a little bit, downgrading some charities to 0.
A note on how I think about criticism
(This was initially meant as part of this post,[1] but while editing I thought it didn't make a lot of sense there, so I pulled it out.)
I came to CEA with a very pro-criticism attitude. My experience there reinforced those views in some ways,[2] but it also left me more attuned to the costs of criticism (or of some pro-criticism attitudes). (For instance, I used to see engaging with all criticism as virtuous, and have changed my mind on that.) My overall takes now aren’t very crisp or easily summarizable, but I figured I'd try to share some notes.
...
It’s generally good for a community’s culture to encourage criticism, but this is more complicated than I used to think.
Here’s a list of things that I believe about criticism:
- Criticism or critical information can be extremely valuable. It can be hard for people to surface criticism (e.g. because they fear repercussions), which means criticism tends to be undersupplied.[3] Requiring critics to present their criticisms in specific ways will likely stifle at least some valuable criticism. It can be hard to get yourself to engage with criticism of your work or things you care about. I
... (read more)A note on mistakes and how we relate to them
(This was initially meant as part of this post[1], but I thought it didn't make a lot of sense there, so I pulled it out.)
“Slow-rolling mistakes” are usually much more important to identify than “point-in-time blunders,”[2] but the latter tend to be more obvious.
When we think about “mistakes”, we usually imagine replying-all when we meant to reply only to the sender, using the wrong input in an analysis, including broken hyperlinks in a piece of media, missing a deadline, etc. I tend to feel pretty horrible when I notice that I've made a mistake like this.
I now think that basically none of my mistakes of this kind — I’ll call them “Point-in-time blunders” — mattered nearly as much as other "mistakes" I've made by doing things like planning my time poorly, delaying for too long on something, setting up poor systems, or focusing on the wrong things.
This second kind of mistake — let’s use the phrase “slow-rolling mistakes” — is harder to catch; I think sometimes I'd identify them by noticing a nagging worry, or by having multiple conversations with someone who disagreed with me (and slowly changing my mind), or by seriously reflec... (read more)
USAID has announced that they've committed $4 million to fighting global lead poisoning!
USAID Administrator Samantha Power also called other donors to action, and announced that USAID will be the first bilateral donor agency to join the Global Alliance to Eliminate Lead Paint. The Center for Global Development (CGD) discusses the implications of the announcement here.
For context, lead poisoning seems to get ~$11-15 million per year right now, and has a huge toll. I'm really excited about this news.
Also, thanks to @ryancbriggs for pointing out that this seems like "a huge win for risky policy change global health effective altruism" and referencing this grant:
In December 2021, GiveWell (or the EA Funds Global Health and Development Fund?) gave a grant to CGD to "to support research into the effects of lead exposure on economic and educational outcomes, and run a working group that will author policy outreach documents and engage with global policymakers." In their writeup, they recorded a 10% "best case" forecast that in two years (by the end of the grant period), "The U.S. government, other international actors (e.g., bilateral and multilateral donors), and/or national ... (read more)
Reflection on my time as a Visiting Fellow at Rethink Priorities this summer
I was a Visiting Fellow at Rethink Priorities this summer. They’re hiring right now, and I have lots of thoughts on my time there, so I figured that I’d share some. I had some misconceptions coming in, and I think I would have benefited from a post like this, so I’m guessing other people might, too. Unfortunately, I don’t have time to write anything in depth for now, so a shortform will have to do.
Fair warning: this shortform is quite personal and one-sided. In particular, when I tried to think of downsides to highlight to make this post fair, few came to mind, so the post is very upsides-heavy. (Linch’s recent post has a lot more on possible negatives about working at RP.) Another disclaimer: I changed in various ways during the summer, including in terms of my preferences and priorities. I think this is good, but there’s also a good chance of some bias (I’m happy with how working at RP went because working at RP transformed me into the kind of person who’s happy with that sort of work, etc.). (See additional disclaimer at the bottom.)
First, some vague background on me, in case it’s relevant:
Thanks for writing this Lizka! I agree with many of the points in this [I was also a visiting fellow on the longtermist team this summer]. I'll throw my two cents in about my own reflections (I broadly share Lizka's experience, so here I just highlight the upsides/downsides things that especially resonated with me, or things unique to my own situation):
Vague background:
Upsides:
Note to onlookers that we at Rethink Priorities will pay up to $2000 for people to upgrade their computers and that we view this as very important! And if you work with us for more than a year, you can keep your new computer forever.
I realize that this policy may not be a great fit for interns / fellows though, so perhaps I will think about how we can approach that.
I think we should maybe just send a new mid-end chromebook + high-end headsets with builtin mic + other computing supplies to all interns as soon as they start (or maybe before), no questions asked. Maybe consider higher end equipment for interns who are working on more compute-intensive stuff and/or if they or their managers asked for it.
For some of the intern projects (most notably on the survey team?), more computing power is needed, but since so much of RP work involves Google docs + looking stuff up fast on the internet + Slack/Google Meet comms, the primary technological bottlenecks that we should try to solve is really fast browsing/typing/videocall latency and quality, which chromebooks and headsets should be sufficient for.
(For logistical reasons I'm assuming that the easiest thing to do is to let the interns keep the chromebook and relevant accessories)
Edit: I've now shared: Donation Election: how voting will work. Really grateful for the discussion on this thread!
We're planning on running a Donation Election for Giving Season.
What do you think the final voting mechanism should be, and why? E.g. approval voting, ranked-choice voting, quadratic voting, etc.
Considerations might include: how well this will allocate funds based on real preferences, how understandable it is to people who are participating in the Donation Election or following it, etc.
I realize that I might be opening a can of worms, but I'm looking forward to reading any comments! I might not have time to respond.
Some context (see also the post):
Users will be able to "pre-vote" (to signal that they're likely to vote for some candidates, and possibly to follow posts about some candidates), for as many candidates as they want. The pre-votes are anonymous (as are final votes), but the total numbers will be shown to everyone. There will be a separate process for final voting, which will determine the three winners in the election. The three winners will receive the winnings from the Donation Election Fund, split proportionally based on the votes.
On... (read more)
I’m a researcher on voting theory, with a focus on voting over how to divide a budget between uses. Sorry I found this post late, so probably things are already decided but I thought I’d add my thoughts. I’m going to assume approval voting as input format.
There is an important high-level decision to make first regarding the objective: do we want to pick charities with the highest support (majoritarian) or do we want to give everyone equal influence on the outcome if possible (proportionality)?
If the answer is “majoritarian”, then the simplest method makes the most sense: give all the money to the charity with the highest approval score. (This maximizes the sum of voter utilities, if you define voter utility to be the amount of money that goes to the charities a voter approves.)
If the answer is “proportionality”, my top recommendation would be to drop the idea of having only 3 winners and not impose a limit, and instead use the Nash Product rule to decide how the money is split [paper, wikipedia]. This rule has a nice interpretation where let’s say there are 100 voters, then every voter is assigned 1/100th of the budget and gets a guarantee that this part is only spent on charities ... (read more)
I think since there can be multiple winners, letting people vote on the ideal distribution then averaging those distributions would be better than direct voting, since it most directly represents "how voters think the funds should be split on average" or similar, which seems like what you want to capture? And also is still very understandable I hope.
E.g. if I think 75% of the pool should go to LTFF and 20% to GiveWell, and 5% to the EA AWF, 0% to all the rest, I vote 75%/20%/5%/0%/0%/0% etc. Then, you take the average of those distributions across all voters. I guess it gets tricky if you are only paying out to the top three, but maybe you can just scale their percentage splits? IDK.
If not that or if it is annoying to implement, IMO approval voting or quadratic are probably best, but am not really sure. Ranked choice feels like it is so explicitly designed for single winner elections that it is harder to apply here.
some thoughts on different mechanisms:
Quadratic voting:
I think this could be fun. An advantage here is that voters have to think about the relative value of different charities, rather than just deciding which are better or worse. This could also be an important aspect when we want people to discuss how they plan to vote/how others should vote. If you want to be explicit about this, you could also consider designing the user interface so that users enter these relative differences of charities directly (e.g. "I vote charity A to be 3 times as good as charity B" rather than "I assign 90 vote credits to charity A and 10 vote credits to charity B"). Note however, that due to the top-3 cutoff, putting in the true relative differences between charities might not be the optimal policy.
A technical remark: If you want only to do payouts for the top three candidates, instead of just relying on the final vote, I think it would be better to rescale the voting credits of each voter after kicking out the charity with the least votes and then repeating the process until there are only 3 charities left. This would reduce tactical voting and would respect voters more who pick unusual charities as ... (read more)
Vasili Arkhipov is discussed less on the EA Forum than Petrov is (see also this thread of less-discussed people). I thought I'd post a quick take describing that incident.
Arkhipov & the submarine B-59’s missile
On October 27, 1962 (during the Cuban Missile Crisis), the Russian diesel-powered submarine B-59 started experiencing[1] nearby depth charges from US forces above them; the submarine had been detected and US ships seemed to be attacking. The submarine’s air conditioning was broken,[2] CO2 levels were rising, and B-59 was out of contact with Moscow. Two of the senior officers on the submarine, thinking that a global war had started, wanted to launch their “secret weapon,” a 10-kiloton nuclear torpedo. The captain, Valentin Savistky, apparently exclaimed: “We’re gonna blast them now! We will die, but we will sink them all — we will not become the shame of the fleet.”
The ship was authorized to launch the torpedo without confirmation from Moscow, but all three senior officers on the ship had to agree.[3] Chief of staff of the flotilla Vasili Arkhipov refused. He convinced Captain Savitsky that the depth charges were signals for the Soviet s... (read more)
If you feel overwhelmed by FTX-collapse-related content on the Forum, you can hide most of it by using a tag filter: hover over the "FTX collapse" tag on the Frontpage (find it to the right of the "Frontpage Posts" header), and click on "Hidden."
[Note: this used to say "FTX crisis," and that might still show up in some places.]
I keep coming back to this map/cartogram. It's just so great.
Superman gets to business [private submission to the Creative Writing Contest from a little while back]
“I don’t understand,” she repeated. “I mean, you’re Superman.”
“Well yes,” said Clark. “That’s exactly why I need your help! I can’t spend my time researching how to prioritize while I should be off answering someone’s call for help.”
“But why prioritize? Can’t you just take the calls as they come?”
Lois clicked “Send” on the email she’d been typing up and rejoined the conversation. “See, we realized that we’ve been too reactive. We were taking calls as they came in without appreciating the enormous potential we had here. It’s amazing that we get to help people who are being attacked, help people who need our help, but we could also make the world safer more proactively, and end up helping even more people, even better, and when we realized that, when that clicked—”
“We couldn’t just ignore it.”
Tina looked back at Clark. “Ok, so what you’re saying is that you want to save people— or help people — and you think there are better and worse ways you could approach that, but you’re not sure which are which, and you realized that instead of rushing off to fight the most immediate threat, yo... (read more)
I recently ran a quick Fermi workshop, and have been asked for notes several times since. I've realized that it's not that hard for me to post them, and it might be relatively useful for someone.
Quick summary of the workshop
Resources
Disclaimers:
***************
***************
***************
***************
***************
************... (read more)
Time-of-perils- or existential-risk-themed image I made with DALL-E:
Moderation updates
Moderation update: We have indefinitely banned 8 accounts[1] that were used by the same user (JamesS) to downvote some posts and comments from Nonlinear and upvote critical content about Nonlinear. Please remember that voting with multiple accounts on the same post or comment is very much against Forum norms.
(Please note that this is separate from the incident described here)
my_bf_is_hot, inverted_maslow, aht_me, emerson_fartz, daddy_of_upvoting, ernst-stueckelberg, gpt-n, jamess
LukeDing (and their associated alt account) has been banned for six months, due to voting & multiple-account-use violations. We believe that they voted on the same comment/post with two accounts more than two hundred times. This includes several instances of using an alt account to vote on their own comments.
This is against our Forum norms on voting and using multiple accounts. We will remove the duplicate votes.
As a reminder, bans affect the user, not the account(s).
If anyone has questions or concerns, please feel free to reach out, and if you think we made a mistake here, you can appeal the decision.
We also want to add:
Moderation update: A new user, Bernd Clemens Huber, recently posted a first post ("All or Nothing: Ethics on Cosmic Scale, Outer Space Treaty, Directed Panspermia, Forwards-Contamination, Technology Assessment, Planetary Protection, (and Fermi's Paradox)") that was a bit hard to make sense of. We hadn't approved the post over the weekend and hadn't processed it yet, when the Forum team got an angry and aggressive email today from the user in question calling the team "dipshits" (and providing a definition of the word) for waiting throughout the weekend.
If the user disagrees with our characterization of the email, they can email us to give permission for us to share the whole thing.
We have decided that this is not a promising start to the user's interactions on the Forum, and have banned them indefinitely. Please let us know if you have concerns, and as a reminder, here are the Forum's norms.
Moderation update:
We have strong reason to believe that Torres (philosophytorres) used a second account to violate their earlier ban. We feel that this means that we cannot trust Torres to follow this forum’s norms, and are banning them for the next 20 years (until 1 October 2042).
Moderation update: We have banned "Richard TK" for 6 months for using a duplicate account to double-vote on the same posts and comments. We’re also banning another account (Anin, now deactivated), which seems to have been used by that same user or by others to amplify those same votes. Please remember that voting with multiple accounts on the same post or comment is very much against Forum norms.
(Please note that this is separate from the incident described here)
We’re issuing [Edit: identifying information redacted] a two-month ban for using multiple accounts to vote on the same posts and comments, and in one instance for commenting in a thread pretending to be two different users. [Edit: the user had a total of 13 double-votes, most far apart and are likely accidental, two upvotes close together on others' posts (which they claim are accidental as well), but two cases of deliberate self upvote from alternative accounts]
This is against the Forum norms around using multiple accounts. Votes are really important for the Forum: they provide feedback to authors and signal to readers what other users found most valuable, so we need to be particularly strict in discouraging this kind of vote manipulation.
A note on timing: the comment mentioned above is 7 months old but went unnoticed at the time, a report for it came in last week and triggered this investigation.
If [Edit: redacted] thinks that this is not right, he can appeal. As a reminder, bans affect the user, not the account.
[Edit: We have retroactively decided to redact the user's name from this early message, and are currently rethinking our policies on the mat... (read more)
Moderation update:
We have strong reason to believe that Charles He used multiple new accounts to violate his earlier 6-month-long ban. We feel that this means that we cannot trust Charles He to follow this forum’s norms, and are banning him from the Forum for the next 10 years (until December 20, 2032).
We have already issued temporary suspensions to several suspected duplicate accounts, including one which violated norms about rudeness and was flagged to us by multiple users. We will be extending the bans for each of these accounts to mirror Charles’s 10-year ban, but are giving the users an opportunity to message us if we have made any of those temporary suspensions in error (and have already reached out to them). While we aren’t >99% certain about any single account, we’re around 99% that at least one of these is Charles He.
You can find more on our rules for pseudonymity and multiple accounts here. If you have any questions or concerns about this, please also feel free to reach out to us at forum-moderation@effectivealtruism.org.
I don't think repeatedly evading moderator bans is a "relatively small crime". If Forum moderation is to mean anything at all, it has to be consistently enforced, and if someone just decides that moderation doesn't apply to them, they shouldn't be allowed to post or comment on the Forum.
Charles only got to his 6 month ban via a series of escalating minor bans, most of which I agreed with. I think he got a lot of slack in his behaviour because he sometimes provided significant value, but sometimes (with insufficient infrequency) behaved in ways that were seriously out of kilter with the goal of a healthy Forum.
I personally think the 10-year thing is kind of silly and he should just have been banned indefinitely at this point, then maybe have the ban reviewed in a little while. But it's clear he's been systematically violating Forum policies in a way that requires serious action.
I have no idea if this was intentional on the part of the moderators, but they aren't all flying in the same direction. ;-)
It's kind of jarring to read that someone has been banned for "violating a norm" - that word to me implies that they're informal agreements between the community. Why not call them "rules"?
pinkfrog (and their associated account) has been banned for 1 month, because they voted multiple times on the same content (with two accounts), including upvoting pinkfrog's comments with their other account. To be a bit more specific, this happened on one day, and there were 12 cases of double-voting in total (which we’ll remove). This is against our Forum norms on voting and using multiple accounts.
As a reminder, bans affect the user, not the account(s).
If anyone has questions or concerns, please feel free to reach out, and if you think we made a mistake here, you can appeal the decision.
Multiple people on the moderation team have conflicts of interest with pinkfrog, so I wanted to clarify our process for resolving this incident. We uncovered the norm violation after an investigation into suspicious voting patterns, and only revealed the user’s identity to part of the team. The moderators who made decisions about how to proceed aren't aware of pinkfrog's real identity (they only saw anonymized information).
Have the moderators come to a view on identifying information? is pinkfrog the account with higher karma or more forum activity?
In other cases the identity has been revealed to various degrees:
LukeDing
JamesS
Richard TK (noting that an alt account in this case, Anin, was also named)
[Redacted]
Charles He
philosophytorres (but identified as "Torres" in the moderator post)
It seems inconsistent to have this info public for some, and redacted for others. I do think it is good public service to have this information public, but am primarily pushing here for consistency and some more visibility around existing decisions.
Agree. It seems potentially pretty damaging to people’s reputations to make this information public (and attached to their names); that strikes me as a much bigger penalty than the bans. There should, at a minimum, be a consistent standard, and I’m inclined to think that standard should be having a high bar for releasing identifying information.
I think we should hesitate to protect people from reputational damage caused by people posting true information about them. Perhaps there's a case to be made when the information is cherry-picked or biased, or there's no opportunity to hear a fair response. But goodness, if we've learned anything from the last 18 months I hope it would include that sharing information about bad behaviour is sometimes a public good.
I suggest editing this comment to note the partial reversal on appeal and/or retracting the comment, to avoid the risk of people seeing only it and reading it as vaguely precedential.
Strong +1 to Richard, this seems a clear incorrect moderation call and I encourage you to reverse it.
I'm personally very strongly opposed to killing people because they eat meat, and the general ethos behind that. I don't feel in the slightest offended or bothered by that post, it's just one in a string of hypothetical questions, and it clearly is not intended as a call to action or to encourage action.
If the EA Forum isn't somewhere where you can ask a perfectly legitimate hypothetical question like that, what are we even doing here?
The moderators have reviewed the decision to ban @dstudioscode after users appealed the decision. Tl;dr: We are revoking the ban, and are instead rate-limiting dstudioscode and warning them to avoid posting content that could be perceived as advocating for major harm or illegal activities. The rate limit is due to dstudiocode’s pattern of engagement on the Forum, not simply because of their most recent post—for more on this, see the “third consideration” listed below.
More details:
Three moderators,[1] none of whom was involved in the original decision to ban dstudiocode, discussed this case.
The first consideration was “Does the cited norm make sense?” For reference, the norm cited in the original ban decision was “Materials advocating major harm or illegal activities, or materials that may be easily perceived as such” (under “What we discourage (and may delete or edit out)” in our “Guide to norms on the Forum”). The panel of three unanimously agreed that having some kind of Forum norm in this vein makes sense.
The second consideration was “Does the post that triggered the ban actually break the cited norm?” For reference, the post ended with the question “sho... (read more)
I appreciate the thought that went into this. I also think that using rate-limits as a tool, instead of bans, is in general a good idea. I continue to strongly disagree with the decisions on a few points:
- I still think including the "materials that may be easily perceived as such" clause has a chilling effect.
- I also remember someone's comment that the things you're calling "norms" are actually rules, and it's a little disingenuous to not call them that; I continue to agree with this.
- The fact that you're not even willing to quote the parts of the post that were objectionable feels like an indication of a mindset that I really disagree with. It's like... treating words as inherently dangerous? Not thinking at all about the use-mention distinction? I mean, here's a quote from the Hamas charter: "There is no solution for the Palestinian question except through Jihad." Clearly this is way way more of an incitement to violence than any quote of dstudiocode's, which you're apparently not willing to quote. (I am deliberately not expressing any opinion about whether the Hamas quote is correct; I'm just quoting them.) What's the difference?
- "They see the fact that it is “just” a philosophical
... (read more)On point 4:
I'm pretty sure we could come up with various individuals and groups of people that some users of this forum would prefer not to exist. There's no clear and unbiased way to decide which of those individuals and groups could be the target of "philosophical questions" about the desirability of murdering them and which could not. Unless we're going to allow the question as applied to any individual or group (which I think is untenable for numerous reasons), the line has to be drawn somewhere. Would it be ethical to get rid of this meddlesome priest should be suspendable or worse (except that the meddlesome priest in question has been dead for over eight hundred years).
And I think drawing the line at we're not going to allow hypotheticals about murdering discernable people[1] is better (and poses less risk of viewpoint suppression) than expecting the mods to somehow devise a rule for when that content will be allowed and consistently apply it. I think the effect of a bright-line no-murder-talk rule on expression of ideas is modest because (1) posters can get much of the same result by posing non-violent scenarios (e.g., leaving someone to drown in a pond is neither an a... (read more)
Ty for the reply; a jumble of responses below.
You are literally talking about the sort of conversations that created EA. If people don't have these conversations on the forum (the single best way to create common knowledge in the EA commmunity), then it will be much harder to course-correct places where fundamental ideas are mistaken. I think your comment proceeds from the implicit assumption that we're broadly right about stuff, and mostly just need to keep our heads down and do the work. I personally think that a version of EA that doesn't have the ability to course-correct in big ways would be net negative for the world. In general it is not possible to e.g. identify ongoing moral catastrophes when you're optimizing your main venue of conversations for avoiding seeming weird.
If you're not able to talk about evil people and their ideologies, then you will not be able to account for them in reasoning about how to steer the world. I think EA is already far ... (read more)
Narrowing in even further on the example you gave, as an illustration: I just had an uncomfortable conversation about age of consent laws literally yesterday with an old friend of mine. Specifically, my friend was advocating that the most important driver of crime is poverty, and I was arguing that it's cultural acceptance of crime. I pointed to age of consent laws varying widely across different countries as evidence that there are some cultures which accept behavior that most westerners think of as deeply immoral (and indeed criminal).
Picturing some responses you might give to this:
- That's not the sort of uncomfortable claim you're worried about
- But many possible continuations of this conversation would in fact have gotten into more controversial territory. E.g. maybe a cultural relativist would defend those other countries having lower age of consent laws. I find cultural relativism kinda crazy (for this and related reasons) but it's a pretty mainstream position.
- I could have made the point in more sensitive ways
- Maybe? But the whole point of the conversation was about ways in which some cultures are better than others. This is inherently going to be a sensitive claim, and it's hard
... (read more)Do you have an example of the kind of early EA conversation which you think was really important which helped came up with core EA tenets might be frowned upon or censored on the forum now? I'm still super dubious about whether leaving out a small number of specific topics really leaves much value on the table.
And I really think conversations can be had in more sensitive ways. In the the case of the original banned post, just as good a philosophical conversation could be had without explicitly talking about killing people. The conversation already was being had on another thread "the meat eater problem"
And as a sidebar yeah I wouldn't have any issue with that above conversation myself because we just have to practically discuss that with donors and internally when providing health care and getting confronted with tricky situations. Also (again sidebar) it's interesting that age of marriage/consent conversations can be where classic left wing cultural relativism and gender safeguarding collide and don't know which way to swing. We've had to ask that question practically in our health centers, to decide who to give family planning to and when to think of referring to police etc. Super tricky.
My point is not that the current EA forum would censor topics that were actually important early EA conversations, because EAs have now been selected for being willing to discuss those topics. My point is that the current forum might censor topics that would be important course-corrections, just as if the rest of society had been moderating early EA conversations, those conversations might have lost important contributions like impartiality between species (controversial: you're saying human lives don't matter very much!), the ineffectiveness of development aid (controversial: you're attacking powerful organizations!), transhumanism (controversial, according to the people who say it's basically eugenics), etc.
Re "conversations can be had in more sensitive ways", I mostly disagree, because of the considerations laid out here: the people who are good at discussing topics sensitively are mostly not the ones who are good at coming up with important novel ideas.
For example, it seems plausible that genetic engineering for human intelligence enhancement is an important and highly neglected intervention. But you had to be pretty disagreeable to bring it into the public conversation a few years ago (I think it's now a bit more mainstream).
This moderation policy seems absurd. The post in question was clearly asking purely hypothetical questions, and wasn't even advocating for any particular answer to the question. May as well ban users for asking whether it's moral to push a man off a bridge to stop a trolley, or ban Peter Singer for his thought experiments about infanticide.
Perhaps dstudiocode has misbehaved in other ways, but this announcement focuses on something that should be clearly within the bounds of acceptable discourse. (In particular, the standard of "content that could be interpreted as X" is a very censorious one, since you now need to cater to a wide range of possible interpretations.)
That is not the post in question. We removed the post that prompted the ban.
Here are slides from my "Writing on the Forum" workshop at EAGxBerlin.
If you voted in the Donation Election, how long did it take you? (What did you spend the most time on?)
I'd be really grateful for quick notes. (You can also private message me if you prefer.)
Tagging posts doesn't work right now — apologies for the inconvenience! The Forum team is working on it, and I hope we'll fix it soon.