Early in a career, if you're uncertain about what you're good at, exploration of your skills/abilities is necessary. Choosing quicker/cheaper tests at this time can help you do this more efficiently.
However, for assessing skill/personal-fit in our priority areas, a lot of the advice we give is "major in X in undergrad" or "be able to get into this type of job/grad school." To my mind, these aren't efficient tests - by the time you've gotten to higher level classes that truly test your ability to potentially move the field forward/get into the right job or grad-school, it's pretty late to pivot. Also, this advice only applies to EAs currently in college.
Instead for priority paths, could/should 80000-Hours and the EA community curate sets of tests, starting from cheap to expensive, that one can use to rapidly gauge their skills? For instance for technical AI safety research, I layout the following hypothetical example (heads up - I'm no AI expert)
- Test 1 - Assess how fundamentally strong/flexible your technical thinking is
- Learn a programming language / fundamentals of programming (ex. MIT's Intro to Python MOOC)
- Learn Data Structures / Algorithms (ex. Princeton's Algorithms MOOC)
- Start doing programming puzzles on Leetcode
- Are you enjoying doing these sorts of computational puzzles?
- Are you able to solve them in a reasonable amount of time? (~30 min)
- If no: you may not have a high likelihood of being a good fit for technical AI safety research.You may be able to contribute to the field in other ways - but you might need to adjust your plans.
- If yes: continue
- Test 2 - (it would go on from here - for instance, now start doing competitive programming and learn fundamentals of Machine Learning)
The advantage of Test 1 is that you've found a way to test the fundamental skill of flexible technical thinking without investing a ton of time just learning accessory information (how vectors/matrices work, how TensorFlow works, etc.). You could arguably figure this out in one summer instead of over many years.The potential downsides are:
- We'd need to make sure the tests truly do serve as a good indicator of the core skill - otherwise we're giving advice that leads to an unacceptable amount of false positives and/or false negatives.
- It can be less motivating to work on proxy problems than learning stuff related to the actual topic of interest, which can throw off the accuracy of the tests.
- We'd have to balance how specific, yet flexible these testing guidelines should be.
 Again, I'm no expert on technical AI research. Feel free to dispute this example if inaccurate, but I'd ask you to try and focus on the broad concept of "could a more accurate set of ranked tests exist and actually be useful for EAs?"