I haven't received anything on my side. I think a confirmation by email would be nice, yes. Otherwise, I'll send the application a second time just in case.
Thanks for writing this Jamie!Concerning the "SHOULD WE FOCUS ON MORAL CIRCLE EXPANSION?" question, I think something like the following sub-question is also relevant: Will MCE lead to a "near miss" of the values we want to spread?
Magnus Vinding (2018) argues that someone who cares about a given sentient being, is absolutely not guaranteed to wish what we think is the best for this sentient being. While he argues from a suffering-focused perspective, the problem is still the same from any ethical framework. For instance, future people who ... (read more)
I completely agree with 3 and it's indeed worth clarifying. Even ignoring this, the possibility of humans being more compassionate than pro-life grabby aliens might actually be an argument against human-driven space colonization, since compassion -- especially when combined with scope sensitivity -- seems to increase agential s-risks related to potential catastrophic cooperation failure between AIs (see e.g., Baumann and Harris 2021, 46:24), which are the most worrying s-risks according to Jesse Clifton's preface of CLR's agenda. A space filled with ... (read more)
Interesting! Thank you for writing this up. :)
It does seem plausible that, by evolutionary forces, biological nonhumans would care about the proliferation of sentient life about as much as humans do, with all the risks of great suffering that entails.
What about the grabby aliens, more specifically? Do they not, in expectation, care about proliferation (even) more than humans do?All else being equal, it seems -- at least to me -- that civilizations with very strong pro-life values (i.e., that thinks that perpetuating life is good and necessary, ... (read more)
Good point, and it is valid with CLR's s-risks definition. :)
Thank you for writing this.
According to a survey of quantitative predictions, disappointing futures appear roughly as likely as existential catastrophes. [More]
It looks like that Bostrom and Ord included risks of disappointing futures in their estimates on x-risks, which might make this conclusion a bit skewed, don't you think?
Michael's definition of risks of disappointing futures doesn't include s-risks though, right?
a disappointing future is when humans do not go extinct and civilization does not collapse or fall into a dystopia, but civilization nonetheless never realizes its potential.
I guess we get something like "risks of negative (or nearly negative) future" adding up the two types.
Great piece, thanks !Since you devoted a subsection to moral circle expansion as a way of reducing s-risks, I guess you consider that its beneficial effects outweigh the backfire risks you mention (at least if MCE is done "in the right way"). CRS' 2020 End-of-Year Fundraiser post also induces optimism regarding the impact of increasing moral consideration for artificial minds (the only remaining doubts seem to be about when and how to do it).I wonder how confident we should be about this (the positiveness of MCE in reducing s-risks), at this point? Have yo... (read more)
Thanks for writing this! :)Another potential outcome that comes to mind regarding such projects is a self-fulfilling prophecy effect (provided the predictions are not secret). I have no idea how much of an (positive/negative) impact it would have though.