As far as I know, there isn't that much funding or research in EA on AI sentience (though there is some? e.g. this)
I can imagine some answers:
- Very intractable
- Alignment is more immediately the core challenge, and widening the focus isn't useful
- Funders have a working view that additional research is unlikely to affect (e.g. that AIs will eventually be sentient?)
- Longtermist focus is on AI as an X-risk, and the main framing there is on avoiding humans being wiped out
But it also seems important and action-relevant:
- Current framing of AI safety is about aligning with humanity, but making AI go well for AI's could be comparably / more important
- Naively, if we knew AIs would be sentient, it might make 'prioritising AIs welfare in AI development' a much higher impact focus area
- It's an example of an area that won't necessarily attract resources / attention from commercial sources
(I'm not at all familiar with the area of AI sentience and posted without much googling, so please excuse any naivety in the question!)
Maybe a question instead of an answer, but what longtermist questions does this seem like a crux for?
If AIs are unaligned with human values, that seems very bad already.
If they are aligned, then surely our future selves can figure this out?
Again, could be very dumb question, but without knowing that, it doesn't seem surprising how little attention is paid to AI sentience.
Scenario 1: Alignment goes well. In this scenario, I agree that our future AI-assisted selves can figure things out, and that pre-alignment AI sentience work will have been wasted effort.
Scenario 2: Alignment goes poorly. While I don’t technically disagree with your statement, “If AIs are unaligned with human values, that seems very bad already,” I do think it misleads through lumping together all kinds of misaligned AI outcomes into “very bad,” when in reality this category ranges across many orders of magnitude of badness.[1] In the case that we los... (read more)