PF

Pedro Freire

@ Independent

Posts
1

Sorted by New

Comments
9

Maybe a more clarifying and charitable title for an 'AI As Normal Technology'-like position would be 'No Major Technological Revolution Has Been Normal'.

Here are some bullet points of reflection topics around lifestyle and priorities for EAs that I shared with some fellow EAs some months ago. I am sharing this text here in case it interests anyone. I will elaborate and expand on them more and better later if I have the opportunity.

""" Support Systems: Seriously. I didn't even know this term until after all this happened, and it would have changed everything. There's something about how people are instructed in STEM institutions (and as a consequence, many EA institutions) that makes it all about careers, how one's impact is understood by their public professional life. And then it turns out that in reality a lot of the most publicly impactful people have these incredibly beautiful family and fraternity systems that were at the core of everything they've done, that never get talked about. Too many yang, public, external, wikipedia-worthy archetypes of impact. It would be really awesome if every youngling EA-in-training knew that having strong and abundant support systems, investing in true family and friends, investing in intimacy, figuring out relationships, being connected to non-EAs... that this sort of thing might be not a distraction from impact but a foundation for impact.

Something something about impact theory: I don't know, there's something about EA theory where it wants it to be really convincing that being an EA is the most important thing to do, but somewhere in all the moral arguments, it takes way too many shortcuts. By taking shortcuts to force it to be the case that being an EA is the right moral thing to do, you are forced to ignore and push under the rug all forms of impact that don't currently fit well into EA career stories and don't have a legible trace of impact connecting it to an EA. I don't really know how to solve this. If I were to give any pointers, here's what first comes to mind:

-- Legibility: there's a serious expectation that impact has to be legible. This is baked into the EA foundation. Unfortunately, in the real world, there are probably more illegible actions of impact than legible ones. Sure, I think we've been adding footnotes on EA material about this, but this is not a small thing that can be addessed separately from the rest of the decision-making. It truly affects the foundation on which the majority of EA arguments are based upon. One has to be able to make decisions in the world incorporating and accepting the fact that the majority of impact is fundamentally illegible, made by people you won't get to know personally, that sometimes public information and public consensus about events can be pretty irrelevant when it comes to understanding and planning on the ground.

-- Argumentation: there's an expectation that truth is found by finding the best arguments. This is true in all cases where this is true, except in all the cases where it isn't. This stems from the above; arguments rely on legible, shared-knowledge facts, and there's just so much of what decides what happens on a daily life basis that is far removed from that. Simplifications are incredibly robust in some cases, and incredibly illusory in others. Obviously we don't want to abandon arguments, but more like, grow beyond them.

-- Curiosity and connection: The majority of good human beings are not EAs! What are they all truly up to?"""

I guess the overall point here is that thinking of saving lives purely in terms of dollars feels like a type error; for it is totally possible today to save lives as an investment, by saving undersupported people from their precarious realities and allowing them to be productive economic workers, eventually generating more revenue than the initial life saving investment.

(My Facebook and Instagram accounts have been suspended without explanation. Hopefully they will be restored soon. If anyone reading this wants to reach me in the meantime, please use other means.)

I agree this cannot replace donation-based interventions! It is still feels potentially underrated and underconsidered.

I do agree that management and structure are the hardest parts. I do imagine many EA orgs have solved harder problems in the past.

I think automatic dubbing services have become good enough to make English fluency not be a hard requirement anymore for many potential jobs.

Here is a super hacky fermi-gpt estimate of a headcount of potentially hireable global workers:

"""

hacky fermi estimate — internet users → elite tail

definitions (clean + explicit):

  • population: total population (≈2024–2025)
  • internet users: people using the internet (any device)
  • final pool (÷8000): internet users filtered by three independent 95th-percentile criteria
    • high cognitive ability (≈95th percentile)
    • hardworking (≈95th percentile)
    • ethical / trustworthy (≈95th percentile)
      combined ⇒ (1 / (20×20×20) ≈ 1 / 8000)

interpretation: this is a very conservative lower bound on people who could plausibly do high-quality remote cognitive work using tools like chatgpt (incl. translation). this is not a hiring claim; it’s an order-of-magnitude sanity check.


hacky fermi table

country population internet users final pool (÷8000)
brazil 203,000,000 170,520,000 21,315
argentina 46,000,000 41,400,000 5,175
colombia 52,000,000 40,040,000 5,005
peru 34,000,000 24,480,000 3,060
chile 19,500,000 17,940,000 2,243
bolivia 12,400,000 7,440,000 930
paraguay 7,500,000 5,850,000 731
ecuador 18,300,000 13,725,000 1,716
mexico 129,000,000 96,750,000 12,094
nigeria 227,000,000 88,530,000 11,066
ghana 34,000,000 18,020,000 2,253
kenya 55,000,000 23,650,000 2,956
uganda 49,000,000 14,210,000 1,776
tanzania 67,000,000 20,100,000 2,513
south africa 62,000,000 44,640,000 5,580
egypt 112,000,000 80,640,000 10,080
morocco 37,000,000 31,080,000 3,885
tunisia 12,300,000 8,733,000 1,092
india 1,430,000,000 800,800,000 100,100
bangladesh 173,000,000 70,930,000 8,866
pakistan 241,000,000 86,760,000 10,845
sri lanka 22,000,000 11,880,000 1,485
vietnam 101,000,000 75,750,000 9,469
philippines 114,000,000 83,220,000 10,403
indonesia 277,000,000 182,820,000 22,853
thailand 71,000,000 60,350,000 7,544
malaysia 34,000,000 32,980,000 4,123
nepal 30,500,000 13,420,000 1,678
cambodia 17,000,000 9,520,000 1,190
mongolia 3,500,000 2,905,000 363
fiji 930,000 697,500 87
samoa 225,000 157,500 20
tonga 107,000 74,900 9

key takeaway:
even after filtering to internet users only and then applying an extremely harsh 95%×95%×95% filter, many countries still have thousands to tens of thousands of plausible high-quality contributors. at global scale, talent supply is not the bottleneck; coordination, tooling, and trust are.

"""

(I know this estimation relies on some independence assumptions. Regardless, it is meant to be illustrative, not authoritative.)

An underexplored alternative to donation is hiring people from low-income contexts to do paid work on meaningful problems.

Here's a rough estimate of "happy" hourly rates for remote intellectual manual labor (data labeling, checking, summarization, interpretability grunt work), in USD:

Estimated Happy Rates ($/h)

Country p25 p50 p75
Brazil 2.35 3.35 4.69
Argentina 2.11 3.02 4.23
Colombia 3.93 5.61 7.85
Peru 2.38 3.40 4.75
Chile 4.75 6.79 9.50
Bolivia 1.70 2.45 3.40
Paraguay 2.05 2.95 4.10
Ecuador 2.70 3.85 5.40
Mexico 2.90 4.10 5.80
Nigeria 0.70 0.99 1.39
Ghana 0.63 0.90 1.26
Kenya 1.24 1.77 2.48
Uganda 0.55 0.80 1.15
Tanzania 0.60 0.88 1.25
South Africa 2.07 2.96 4.14
Egypt 1.46 2.09 2.92
Morocco 1.85 2.65 3.70
Tunisia 1.95 2.80 3.90
India 0.95 1.40 2.10
Bangladesh 0.55 0.80 1.20
Pakistan 0.65 0.95 1.40
Sri Lanka 0.85 1.25 1.85
Vietnam 1.35 1.95 2.80
Philippines 1.60 2.30 3.30
Indonesia 1.10 1.60 2.40
Thailand 2.10 3.00 4.30
Malaysia 2.60 3.70 5.30
Nepal 0.60 0.88 1.30
Cambodia 0.75 1.10 1.60
Mongolia 1.10 1.60 2.30
Fiji 2.40 3.40 4.90
Samoa 2.10 3.00 4.30
Tonga 2.20 3.10 4.50

There exists a very large supply of people who are both willing and happy to do careful cognitive work at rates that are trivial by rich-country standards, if the work is structured and paid.

Some reasons this possibility can be quite good and interesting:

  • It allows money to be converted into actual work on impactful tasks, even if that work is initially "intellectual manual labor" (labeling, checking, summarizing, auditing, interpretability grunt work, etc.).
  • It treats people as participants rather than recipients. Receiving payment for work tends to be more humanizing than receiving aid, because it encodes agency, skill, reciprocity, and contribution.
  • It onboards people into the global intellectual labor market: deadlines, quality standards, tooling, communication norms. Those skills compound and transfer.
  • It can operate without heavy intermediary organizations, which reduces overhead and incentive distortion and keeps the causal chain legible: money → work → output → learning.

A lot of important research and analysis is not bottlenecked on genius so much as on coordination, paradigms, and tooling. Once those exist, large amounts of careful human attention can be usefully applied in parallel.

My usual joke is "GPT-2 has 12 attention heads per layer and 48 layers. If we had 50 interns and gave them each a different attention head every day, we'd have an intern-day of analysis of each attention head in 11 days."

This is bottlenecked on various things:

  • having a good operationalization of what it means to interpret an attention head, and having some way to do quality analysis of explanations produced by the interns. This could also be phrased as "having more of a paradigm for interpretability work".
  • having organizational structures that would make this work
  • building various interpretability tools to make it so that it's relatively easy to do this work

Buck's comment on "How might a herd of interns help with AI or biosecurity research tasks/questions?", EA Forum

https://www.encultured.ai/ might be somewhere of your interest? i'd be curious to hear what they think