Background: I'm an undergraduate CS major. Recently, I've mentioned to my mom that I've been getting involved in the "effective altruism" community, and I've been expressing an increased interest in getting a PhD. The other day, my mom asked me why exactly I wanted a PhD.
Me: Well, I want to help others as much as possible.
Mom: Okay, how are you going to help people with a PhD?
Me: Well, I don't know... maybe try to reduce existential risks...
Mom: Whoa, existential risks?
Me: Uh, I don't know, I mean, maybe it wouldn't be that bad, but it seems likely that AI will be very important in the future. And if AI has good goals that match up with the goals of humans, they could solve lots of the world's problems, so I really want to increase the odds of that happening.
Mom: So what's going to happen if AIs don't have good goals?
Me: Well, I guess... they could kill off humanity?
Mom: Whoa!
Fortunately, we moved on in the conversation at this point, but I don't think I gave her the best first impression of these ideas. Does anyone know of any good articles or videos for a popular audience that present the AI alignment problem in moderate depth, without too much sensationalism? I'm sure there are people who would do a much better job than me at explaining these concepts to my mom. Similarly, content on EA concepts in general would be helpful.
It's most important to me to convince my mom that what I'm doing is worthwhile, but I also want to be able to talk about my career plans with non-EAs without them thinking I've joined a Doomsday cult. For people working in existential risk and other "weird" areas - how do you usually talk about your work when it comes up in conversation?
Perhaps try explaining by analogy, or providing examples of ways we’re already messing up.
Like the YouTube algorithm. It only maximizes the amount of time people spend on the platform, because (charitably) Google thought that’d be a useful metric for the quality of the content it provides. But instead, it ended up figuring out that if it showed people videos which convinced them of extreme political ideologies, then it would be easier to find videos which would make them angry/happy/sad/other addictive emotions which would keep them on the platform.
This particular problem has since been fixed, but it took quite a while to figure out what was going on, and more time to figure out how to fix it. Maybe use analogies of genies who, if you imperfectly specify your wish, will find some way to technically satisfy it, but screw you over in the process.
One thing which stops me from explaining things well to my parents is the fear of looking weird. Which usually doesn’t stop me (to a fault) when talking with anyone else, but I guess not with my parents. You can avert this via ye-olde Appeal to Authority. Tell them the idea was popularized, in part, by professor Stuart Russel—the writer of the world’s foremost textbook on artificial intelligence—in his book Human Compatible, who currently runs the organization HCAI at Berkeley to tackle this very problem.
edit: Also, be sure to note it’s not just HCAI who’s working on this problem. There’s also MIRI, DeepMind, Anthropic, and other organizations.