According to public reports, Dan Hendrycks has been influenced by EA since he was a freshman (https://www.bostonglobe.com/2023/07/06/opinion/ai-safety-human-extinction-dan-hendrycks-cais/).
He did the 80,000 hours program.
He worries about AI bringing about the end of humanity, if not the planet.
After getting his Ph.D., he started an AI safety organization instead of joining one of the many AI startups.
And he's taken $13M in donations from two EA orgs - OpenPhilanthropy and FTX Foundation.
Yet he denies being an Effective Altruism member when asked about it by the press. For instance (https://www.bloomberg.com/news/newsletters/2024-06-27/an-up-and-coming-ai-safety-thinker-on-why-you-should-still-be-worried)
As an aside, Hendrycks is not alone in this. The founders of the Future of Life Institute have done the same thing (https://www.insidecyberwarfare.com/p/an-open-source-investigation-into).
I'm curious to know what others think about Hendryck's attempts to disassociate himself from Effective Altruism.
I agree with some of this comment but I really don't get the link to the paper you linked:
The paper seems to mostly be evidence that the benchmarks that, you and other who have been focused on certain kinds of ML experiments have created, are not really helping much with AI alignment.
I also disagree some with the methodology of this paper, but I have trouble seeing how its evidence of people doing too much armchair analyzing, when as far as I can tell the flaws with these benchmarks were the result of people doing too much "IDK what alignment is, but maybe if we measure this vaguely related thing it will help" and too little "man, I should really understand what I would learn if this benchmark improved and whether it would cause me to actually update the system that has improved on this benchmark is more aligned and less likely to cause catastrophic consequences".