Hello! 

I’m Toby, the new Content Manager @ CEA. 

Before working at CEA, I studied Philosophy at the University of Warwick, and worked for a couple of years on a range of writing and editing projects in the EA space. Recently I helped run the Amplify Creative Grants program, in order to encourage more impactful podcasting and YouTube projects (such as the podcast in this Forum post). You can find a bit of my own creative output on my more-handwavey-than-the-ea-forum blog, and my (now inactive) podcast feed.

I’ll be doing some combination of: moderating, running events on the Forum, making changes to the Forum based on user feedback, writing announcements, writing the Forum Digest and/or the EA Newsletter, participating in the Forum a lot etc… I’ll be doubling the capacity of the content team (the team formerly known as Lizka). 

I’m here because the Forum is great in itself, and safeguards parts of EA culture I care about preserving. The Forum is the first place I found online where people would respond to what I wrote and actually understand it. Often they understood it better than I did. They wanted to help me (and each other) understand the content better. They actually cared about there being an answer. 

The EA community is uniquely committed to thinking seriously about how to do good. The Forum does a lot to maintain that commitment, by platforming critiques, encouraging careful, high-context conversations, and sharing relevant information. I’m excited that I get to be a part of sustaining and improving this space. 

I’d love to hear more about why you value the Forum in the comments (or, alternatively, anything we could work on to make it better!)

This is the image I'm using for my profile picture. It's a linoprint I made of one of my favourite statues, The Rites of Dionysus.


 

81

0
0
10

Reactions

0
0
10
Comments23


Sorted by Click to highlight new comments since:

Just to be clear, Lizka isn't being replaced and you're a new, additional content manager? Or does Lizka have a new role now?

Yep, Lizka is still Content Specialist, and I'm additive. There were a lot of great content related ideas being left on the table because Lizka can't do everything at once. So once I'm up to speed we should be able to get even more projects done. 

What's the difference between a Content Specialist and a Content Manager?

The difference in role titles reflects the fact that Lizka is the team lead (of our team of two). From what I understand, the titles needn't make much difference in practice.

PS- I'm presuming there is a disagree react on my above comment because Lizka can in fact do everything at once. Fair enough. 

FWIW I would've expected the Content Manager manages the Content Specialist, not the other way around.

FWIW I would have guessed the reverse re role titles

Yes I am also curious about the difference. I’ve been using them interchangeably.

(I'd guess the different titles mostly just reflect the difference in seniority? cf. "program officer" vs "program associate")

Wow, seeing as HILTS is hands down my favorite podcast so now I’m quite excited to see what new and exciting content will come from the forum. Welcome to the EA Forum team!

Thank you Constance! I'm glad to hear you like the podcast. To be very clear- everything you like about the podcast is down to James and Amy, we just chose to fund them. 

The only thing that comes to mind for me regarding "make it better" would be to change the wording on the tooltips for voting to clarify (or to police?) what they are for. I somewhat regularly see people agree vote or disagree vote with comments that don't contain any claims or arguments.

Interesting! Let me know if any examples come up (feel free to post here or dm). Ideally we wouldn't have the disagree button playing the same role as the karma button. 

Sure. The silly and simplified cliché is something like this: a comment describes someone's feelings (or internal state) and then gets some agree votes and disagree votes, as if Person A says "this makes me happy" and person be wants to argue that point.

(to be clear, this is a very small flaw/issue with the EA Forum, and I wouldn't really object if the people running the forum decide that this is too minor of an issue to spend time on)

A few little examples:

  • Peter Wildeford's comment on this post "What's the difference between a Content Specialist and a Content Manager?" currently has two agree votes. There isn't any argument or stance there; it is merely asking a question. So I assume people are using the agree vote to indicate something like "I also have this question" or "I am glad that you are asking this question."
  • I made a comment a few days ago about being glad that I am not the only one who wants to have financial runway before donating. It currently has a few agree votes and disagree votes, and I can't for the life of me figure out why. There aren't really any stances or claims being made in that comment.
  • Ben West made a comment about lab grown meat that currently has 27 agree votes, even through the comment has nothing to agree with: "Congratulations to Upside Foods, Good Meat, and everyone who worked on this technology!" I guess that people are using the agree vote to indicate something like "I like this, and I want to express the same gratitude."

Is this a problem? Seems fine to me, because the meaning is often clear, as in two of your examples, and I think it adds value in those contexts. And if it's not clear, doesn't seem like a big loss compared to a counterfactual of having none of these types of vote available.

Thanks for putting these together. This doesn't currently seem obviously bad to me for (I think) the same reasons as Isaac Dunn (those examples don't show valueless reacts, and most cases are much clearer). However, your cases are interesting. 

I agree with your read of the reactions to Ben West's comment. 

In the question about my role, perhaps it is slightly less clear, because "I agree that this is a good question" or "I have this question as well" could probably be adequately expressed with Karma. But I also doubt that this has led to significant confusion. 

In the reaction to your comment, I'd go with the agrees saying that they echo the statement in your tl;dr. The disagree is weirder- perhaps they are signalling disencouragement of your encouraging Lizka's sentiment? 


(Perhaps how perplexing people find agree/disagree reacts to comments which don't straightforwardly contain propositions maps to how habitually the reader decouples propositional content from context.) 


I'll keep an eye out for issues with this- my view is loosely held. Thanks again for raising the issue. 
 

Congratulations on the new role! :)

Welcome! Glad to have you here, Toby.

Thanks Joseph!

Welcome Toby :)

Thank you Max!

Congrats Toby, excited to see what you get up to in the new role! And thanks for all your work on Amplify.

Curated and popular this week
Paul Present
 ·  · 28m read
 · 
Note: I am not a malaria expert. This is my best-faith attempt at answering a question that was bothering me, but this field is a large and complex field, and I’ve almost certainly misunderstood something somewhere along the way. Summary While the world made incredible progress in reducing malaria cases from 2000 to 2015, the past 10 years have seen malaria cases stop declining and start rising. I investigated potential reasons behind this increase through reading the existing literature and looking at publicly available data, and I identified three key factors explaining the rise: 1. Population Growth: Africa's population has increased by approximately 75% since 2000. This alone explains most of the increase in absolute case numbers, while cases per capita have remained relatively flat since 2015. 2. Stagnant Funding: After rapid growth starting in 2000, funding for malaria prevention plateaued around 2010. 3. Insecticide Resistance: Mosquitoes have become increasingly resistant to the insecticides used in bednets over the past 20 years. This has made older models of bednets less effective, although they still have some effect. Newer models of bednets developed in response to insecticide resistance are more effective but still not widely deployed.  I very crudely estimate that without any of these factors, there would be 55% fewer malaria cases in the world than what we see today. I think all three of these factors are roughly equally important in explaining the difference.  Alternative explanations like removal of PFAS, climate change, or invasive mosquito species don't appear to be major contributors.  Overall this investigation made me more convinced that bednets are an effective global health intervention.  Introduction In 2015, malaria rates were down, and EAs were celebrating. Giving What We Can posted this incredible gif showing the decrease in malaria cases across Africa since 2000: Giving What We Can said that > The reduction in malaria has be
Ronen Bar
 ·  · 10m read
 · 
"Part one of our challenge is to solve the technical alignment problem, and that’s what everybody focuses on, but part two is: to whose values do you align the system once you’re capable of doing that, and that may turn out to be an even harder problem", Sam Altman, OpenAI CEO (Link).  In this post, I argue that: 1. "To whose values do you align the system" is a critically neglected space I termed “Moral Alignment.” Only a few organizations work for non-humans in this field, with a total budget of 4-5 million USD (not accounting for academic work). The scale of this space couldn’t be any bigger - the intersection between the most revolutionary technology ever and all sentient beings. While tractability remains uncertain, there is some promising positive evidence (See “The Tractability Open Question” section). 2. Given the first point, our movement must attract more resources, talent, and funding to address it. The goal is to value align AI with caring about all sentient beings: humans, animals, and potential future digital minds. In other words, I argue we should invest much more in promoting a sentient-centric AI. The problem What is Moral Alignment? AI alignment focuses on ensuring AI systems act according to human intentions, emphasizing controllability and corrigibility (adaptability to changing human preferences). However, traditional alignment often ignores the ethical implications for all sentient beings. Moral Alignment, as part of the broader AI alignment and AI safety spaces, is a field focused on the values we aim to instill in AI. I argue that our goal should be to ensure AI is a positive force for all sentient beings. Currently, as far as I know, no overarching organization, terms, or community unifies Moral Alignment (MA) as a field with a clear umbrella identity. While specific groups focus individually on animals, humans, or digital minds, such as AI for Animals, which does excellent community-building work around AI and animal welfare while
Recent opportunities in Building effective altruism
46
Ivan Burduk
· · 2m read