Hey everyone, I am extremely excited to write my first ever post on the EA forum!

I figure, why not come out swinging and go Meta right off the bat with a post asking which of my post ideas I should create a post about? After all, the meta, the better (said with a British accent,) right?

I only joined the movement three months ago with the University of Southern California group, but have been thinking along EA lines for many years. I have collected a couple dozen post ideas as I’ve been learning about EA in most of my free time over the past few months. I’d love any feedback on which would be most interesting and useful to community members! I’d also appreciate references to work already written on these topics.

  1. If AI inevitably or almost inevitably dominates the future of the universe, then the hard problem of consciousness and how to ensure conscious, happiness-seeking AI may be the most important cause area

  2. Are intelligence, motivation, consciousness, and creativity all orthogonal, especially at the upper limit of each? If not, what does this mean for AI?

  3. An analysis of Impact Investing as an EA cause area

  4. Analysis of the identity of consciousness, i.e. is consciousness instantaneous (as in Buddhism, non-self or emptiness), continuous over a lifetime (similar to the notion of a soul, with your consciousness starting at birth and ending at death), or universal (the exact same consciousness is in every conscious being simultaneously); AND what does this mean for practical ethics and the long-term future of the universe

  5. Could the future state of a democratic, believability weighted public goods market, essentially a futarchy system, be the system which Yudkowsky’s coherent extrapolated volition, an AI alignment mechanism, uses to have AI predict humanity’s ultimate preferences?

  6. Why I’m more concerned about human alignment than AI alignment; why rapidly accelerating technology will make terrorism an insurmountable existential threat within a relatively short timeframe

  7. Dramatic and easy IQ boost for EAs; evidence suggests creatine boosts IQ by 15 points in vegans. And the importance of vegan supplements generally

  8. Social psychology forces which may cause EAs to be hyper-focused on AI

  9. Massively scalable project-based community building idea

  10. Takeaways from EAGx Boston

  11. Why existntial hope (positive longtermism) may be much more effective at reducing existential risk than trying to reduce existential risk

  12. Speeding up human moral development may be the most effective animal welfare intervention

  13. A series on effective entrepreneurship

  14. My approach to organizational design

  15. Marketing survey on what EA messaging has been most persuasive to comminty members

  16. Why I think broad longtermism is massively underrated

  17. What if we had a perpetual donor contest and entrepreneurial ecosystem rather than just a donor lottery?

  18. The joys of Blinkist (book summary app) for rapid broad learning

  19. Is there a GiveWell for longtermism? There should be.

  20. My current possible trajectories and request for feedback/career advice

  21. How I came to longtermism on my own, and what I think EA longtermism may be getting wrong

  22. Initial thoughts on creating a broad longtermism fellowship

  23. EA dating site idea and prototype

  24. Ultimate Pleasure Machine Dilemma: If you had the opportunity to press a button that turns the entire universe into a perpetual motion pleasure machine, which eternally forces the entire universe into a state of maximum happiness (however you define that), would you press it? (This one was inspired by USC EA Strad Slater)

Feel free to just comment the number or numbers you think is most effective, or to argue why you think so. Really appreciate your feedback, thanks everyone!

26

6 comments, sorted by Click to highlight new comments since: Today at 10:58 PM
New Comment
9. Massively scalable project-based community building idea

If your idea for this is good this might be the highest value post you could write from this list.

20 and 21 (before you get too familiar with EA thinking and possibly forget your origin story) also seem high value.

If 17 is a novel practical idea it's probably also worth writing about.

8 and 16 interest me.

Thanks William! This feedback is super valuable. Yes I think the massive scalable community building project would be novel and it actually ties in with the donor contest as well. Glad to know this would be useful! And good thought, I think writing about my own story will be easiest as well. And I will definitely write about broad longtermism, it is one of my main areas of interest.

6. Why I’m more concerned about human alignment than AI alignment; why rapidly accelerating technology will make terrorism an insurmountable existential threat within a relatively short timeframe

I was thinking about the human alignment portion of this earlier today--how bad actors with future powerful (non-AGI) AI systems at their disposal could cause a tremendous amount of damage. I haven't thought through just how severe this damage might get and would be interested in reading your thoughts on this. What are the most significant risks from unaligned humans empowered by future technology?

Yes! I think the main threats are hard to predict, but mostly involve terrorism with advanced technology, for example weaponized blackholes, intentional grey goo, super coordinated nuclear attacks, and probably many, many other hyper-advanced technilogies we can’t even conceive of yet. I think if technology continues to accellerate it could get pretty bad pretty fast, and even if we’re wrong about AI somehow, human malevolence will be a massive challenge.

Hey Jordan! Great to see another USC person here. The best writing advice I've gotten (that I have yet to implement) is to identify a theory of change for each potential piece--something to keep in mind!

6 sounds interesting, if you can make a strong case for it. Aligning humans isn't an easy task (as most parents, employers, governments, and activists know very well), so I'm curious to hear if you have tractable proposals.

7 sounds important given that a decent number of EAs are vegan, and I'm quite surprised I haven't heard of this before. 15 IQ points is a whole standard deviation, so I'd love to see the evidence for that.

8 might be interesting. I suspect most people are already aware of groupthink, but it could be good to be aware of other relevant phenomena that might not be as widely-known (if there are any).

From what I can tell, 11 proposes a somewhat major reconsideration of how we should approach improving the long-term future. If you have a good argument, I'm always in favor of more people challenging the EA community's current approach. I'm interested in 21 for the same reason.

(In my experience, the answer to 19 is no, probably because there isn't a clear, easy-to-calculate metric to use for longtermist projects in the way that GiveWell uses cost-effectiveness estimates.)

Out of all of these, I think you could whip up a draft post for 7 pretty quickly, and I'd be interested to read it!

Dang yeah I did a quick search on creatine and the IQ number right before writing this post, but now it’s looking like that source was not credible. Would have to research more to see if I can find an accurate reliable measure of creatine cognitive improvement, it seems it at least has a significant impact on memory. Anecdotally, I noticed quite a difference when I took a number of supplements while vegan, and I know there’s some research on a number of differences of various nutrients which vegans lack related to cognitive function. Will do a short post on sometime!

I think human alignment is incredibly difficult, but too important to ignore. I have thought about it a very long time so do have some very ambitious ideas that could feasibly start small and scale up.

Yes! I have been very surprised since joining how narrowly longtermism is focused. I think if the community is right about AGI being within a few decades with fast takeoff then broad longtermism may be less appealing, but I think if there is any doubt about this then we are massively underinvested in broad longtermism and putting all eggs in one basket so to speak. Will definitely write more about this!

Right, definitely wouldn’t be exactly analogous to GiveWell, but I think nonetheless it is important to have SOME way of comparing all the longtermist projects to know what a good investment looks like.

Thanks again for all the feedback Aman! Really appreciate it (and everything else you do for the USC group!!) and really excited to write more on some of these topics :)