All of Diego_Caleiro's Comments + Replies

This is a little old, but it's a similar concept with far higher level of investment: https://www.lesswrong.com/posts/GfRKvER8PWcMj6bbM/sidekick-matchmaking

I haven't read the whole thing. But this seems to be one of the, if not the coolest idea in EA in 2018. Glad you did it.

Good luck for everyone who goes to live or work there!

It has been about 3 years, and only very specific talent still matters for EA now. Earning to Give to institutions is gone, only giving to individuals still makes sense.

It is possible that there will be full scale repleaceability of non-researchers in EA related fields by 2020.

But only if, until then, we keep doing things!

Kaj, I tend to promote your stuff a fair amount to end the inferential silence, and it goes without saying that I agree with all else you said.

Don't give up on your ideas or approach. I am dispirited that there are so few people thinking like you do out there.

It's been less than two years and all the gaps have either been closed, or been kept open in purpose, which Ben Hoffman has been staunchly criticising.

But anyway, it has been less than 2 years and Open Phil has way more money than it knows what to do with.

QED.

-1
Diego_Caleiro
It has been about 3 years, and only very specific talent still matters for EA now. Earning to Give to institutions is gone, only giving to individuals still makes sense. It is possible that there will be full scale repleaceability of non-researchers in EA related fields by 2020. But only if, until then, we keep doing things!

Amanda Askell has interesting thoughts suggestive of using "care" to have a counterfactual meaning. She suggests we think of care as what you would have cared about if you were in a context such that this was a thing you could potentially change. In a way, the distinction is between people who think about "care" in terms of rank "oh, that isn't the thing I most care about" and those who care in terms of absolutes "oh, I think the moral value of this is positive." further complicated by the fact some people are thin... (read more)

0
xccf
Eliezer recently posted advice for central banks on how to accelerate economic growth. I'm not sure if that means he has changed his mind. (Maybe he's deliberately giving them bad advice.)

They need not imply, but I would like a framework where they do under ideal circumstances. In that framework - which I paraphrase from Lewis - if I know a certain moral fact, e.g., that something is one of my fundamental values, then I will value it (this wouldn’t obtain if you are a hypocrite, in such case it wouldn’t be knowledge).

I should X = A/The moral function connects my potential actions to set X. I think I should X = The convolution of the moral function and my prudential function take my potential actions to set X.

I’m unsure I got your notat

... (read more)

I find the idea that there are valid reasons to act that are not moral reasons weird; I think some folks call them prudential reasons. It seems that your reason to be an EA is a moral reason if utilitarianism is right, and “just a reason” if it isn't. But if not what is your reason for doing it?

My understanding of prudential reasons is that they are reasons of the same class as those I have to want to live when someone points a gun at me. They are reasons that relate me to my own preferences and survival, not as a recipient of the utilitarian good, but... (read more)

1
joaolkf
That seems about right under some moral theories. I would not want to distinguish being the recipient of the utilitarian good and getting back massages. I would want to say getting back massages instantiate the utilitarian good. According to this framework, the only thing these prudential reasons capture not in impersonal reasons themselves is the fact people give more weight to themselves than others, but I would like to argue there are impersonal reasons for allowing them to do so. If that fails, then I would call these prudential reasons pure personal reasons, but I would not remove them from the realm of moral reasons. There seems to be already established moral philosophers that tinker with apparently similar types of solutions. (I do stress the “apparently” given that I have not read them fully or fully understand what I read.) They need not imply, but I would like a framework where they do under ideal circumstances. In that framework - which I paraphrase from Lewis - if I know a certain moral fact, e.g., that something is one of my fundamental values, then I will value it (this wouldn’t obtain if you are a hypocrite, in such case it wouldn’t be knowledge). If I value it, and if I desire as I desire to desire (which wouldn’t obtain in moral akrasia), then I will desire it. If I desire it, and if this desire is not outweighed by other conflicting desires (either due to low-level desire multiplicity or high-level moral uncertainty), and if I have moral reasoning to do what servers my desires according to my beliefs (wouldn't obtain for a psychopath), then I will pursue it. And if my relevant beliefs are near enough true, then I will pursue it as effectively as possible. I concede valuing something may not lead to pursuing it, but only if something goes wrong in this chain of deductions. Further, I claim this chain defines what value is. I’m unsure I got your notation. =/= means different? What is the meaning of “/” in “A/The…”? I would claim you are mistaken

I really appreciate your point about intersubjective tractability. It enters the question of how much should we let empirical and practical considerations spill into our moral preferences (ought implies can for example, does it also imply "in a not extremely hard to coordinate way"?)

At large I'd say that you are talking about how to be an agenty Moral agent. I'm not sure morality requires being agenty, but it certainly benefits from it.

Bias dedication intensity: I meant something ortogonal to optimality. Dedicating only to moral preferences, bu... (read more)

Suggestion: Let people talk about any accomplishments, without special emphasis on the month level, or the name of the month.

Some of the moments when people most need to brag is when they need to recover a sense of identity with a self that is more than a month old, that did awesome stuff.

Example: Once upon a time 12 years ago I thought the most good I could do was fixing aging, so I found Aubrey, worked for them for a bit, and won a prize!

A thing I'm proud off is that a few days ago I gave an impromptu speech at Sproul Hall (where free speech started) at Berkeley, about technological improvement and EA, and several people came after to thank me for it.

0
Gleb_T
Meta - Nice suggestion, makes sense. Thanks! Cool, great to hear about the speech!

I frequently use surnames, but in this case since it was a call to action of sorts, first names seemed more appropriate. Thanks for the feedback though, makes sense.

Agreed with 2 first paragraphs.

Activities that are more moral than EA for me: At the moment I think working directly on assembling and conveying knowledge in philosophy and psychology to the AI safety community has higher expected value. I'm taking the AI human compatible course at Berkeley, with Stuart Russell, I hang out at MIRI a lot, so in theory I'm in good position to do that research and some of the time I work on it. But I don't work on it all the time, I would if I got funding for our proposal.

But actually I was referring to a counterfactual wor... (read more)

Telofy: Trying to figure out the direction of the inferential gap here. Let me try to explain, I don't promise to succeed.

Aggregative consequentialist utilitarianism holds that people in general should value most minds having the times of their lives, where "in general" here actually translated into a "should" operator. A moral operator. There's a distinction between me wanting X, and morality suggesting, requiring, or demanding X. Even if X is the same, different things can hold a relation to it.

At the moment I both hold a personal p... (read more)

0
Dawn Drescher
Thanks for bridging the gap! Yeah, that is my current perspective, and I’ve found no meaningful distinction that would allow me to distinguish moral from amoral preferences. What you call intersubjective is something that I consider a strategic concern that follows from wanting to realize my moral preferences. I’ve wondered whether I should count the implications of these strategic concerns into my moral category, but that seemed less parsimonious to me. I’m wary of subjective things and want to keep them contained the same way I want to keep some ugly copypasted code contained, black-boxed in a separate module, so it has no effects on the rest of the code base. I like to use two different words here to make the distinction clearer, moral preferences and moral goals. In both cases you can talk about instrumental and terminal moral preferences/goals. This is how I prefer to distinguish goals from preferences (copypaste from my thesis): To aid comprehension, however, I will make an artificial distinction of moral preferences and moral goals that becomes meaningful in the case of agent-relative preferences: two people with a personal profit motive share the same preference for profit but their goals are different ones since they are different agents. If they also share the agent-neutral preference for minimizing global suffering, then they also share the same goal of reducing it. I’ll assume that in this case we’re talking about agent-neutral preferences, so I’ll just use goal here for clarity. If someone has the personal goal of wanting to get good at playing the theremin, then on Tuesday morning, when they’re still groggy from a night of coding and all out of coffee and Modafinil, they’ll want to stay in bed and very much not want to practice the theremin on one level but still want to practice the theremin on another level, a system 2 level, because they know that to become good at it, they’ll need to practice regularly. Here having practiced is an instrumental

It seems that you feel the moral obligation strongly from your comment. Like the Oxford student cited by Krishna you don't want to do what you want to do, you want to do what you oughtto do.

I don't experience that feeling, so let me reply to your questions:

Wouldn't virtue ethics winning be contradicted by your pulling the lever?

Not really, the pulling of the lever is what I would do, it is what I would think I have reason to do, but it is not what I think I would have moral reason to do. I would reason that a virtuous person (ex hypothesi) wouldn't ... (read more)

I'm not claiming this is optimal, but I might be claiming what I'm about to say may be more optimal than anything else that 98% of EAs are actually doing.

There are a couple thousand billionaires on the planet. There are also about as many EAs.

Let's say 500 billionaires are EA friendly under some set of conditions. Then it may well be that the best use of the top 500 EAs is to minutiously study single individual billionaires. Understand their values, where they come from, what makes them tick. Draw their CT-chart, find out their attachment style, persona... (read more)

As Luke and Nate would tell you, the shift from researcher to CEO is a hard one to make, even when you want to do good, as Hanson puts it "Yes, thinking is indeed more fun."

I have directed an institute in Brazil before, and that was already somewhat a burden.

The main reason for the high variance though is that setting up an institute requires substantial funding. The people most likely to fundraise would be me, Stephen Frey (who is not on the website), and Daniel, and fundraising is taxing in many ways. Would be great if we had for instance the... (read more)

1) Convergence Analysis: The idea here is to create a Berkeley affiliated research institute that operates mainly in two fronts 1)Strategy on the long term future 2)Finding Crucial Considerations that have not been considered or researched yet. We have an interesting group of academics and I would take a mixed position of CEO and researcher.

2) Altruism: past, present, propagation: this is a book whose table of contents I already wrote, and would need further research and spelling out each of the 250 sections I have in mind. It is very different in nature ... (read more)

0
MikeJohnson
All of these seem potentially valuable. I suspect the best choice is the one you'll be most motivated to pursue. My suggestion is that you should consider who your 'customers' are for each project, and figure out which group you'd most like to work with and generate deliverables for. Also, some of these may lend themselves better to intermediate/incremental deliverables, which would be a big plus. ---------------------------------------- All of the above is fully general advice- my low-resolution take on your specific situation is that Convergence Analysis seems by far both the highest leverage (and certainly the largest variance), though the fact that you seem unsure whether to dive down that path may imply there may be some difficult hurdles or complications down that path that you're dreading?
1
Brian_Tomasik
Thanks for sharing. :) If there's a way to encourage Russell to write or teach a bit more about AI safety (even just in his textbook, or maybe in other domains), I would think that would be quite important. But you probably have a better picture of how (in)feasible that is. Sorry that I don't have strong opinions on the other options....
3
RomeoStevens
Three of your projects rank highest on personal interest. I think I would attempt a more granular analysis of this keeping in mind your current uncertainty about your model of your maintained future interest. Some ideas: Pretend that you are handing off the project to someone else and writing a short guide to what you think their curriculum of research and work will look like. Brainstorm the key assumptions underlying the model that assumes a value for each project and see if any of those key assumptions are very cheaply testable (this can be a surprising exercise IME) Premortem (murphyjitsu) each project and compare the likelihood of different failure modes.

Ok, so this doubles as an open thread?

I would like some light from the EA hivemind. For a while now I have been mostly undecided about what to do with my 2016-2017 period.

Roxanne and I even created a spreadsheet so I could evaluate my potential projects and drop most of them, mid-2015. My goals are basically an oscillating mixture of

1)Making the world better by the most effective means possible.

2)Continuing to live in Berkeley

3)Receive more funding

4)Not stop PHD

5)Use my knowledge and background to do (1).

This has proven an extremely hard decision t... (read more)

6
Diego_Caleiro
1) Convergence Analysis: The idea here is to create a Berkeley affiliated research institute that operates mainly in two fronts 1)Strategy on the long term future 2)Finding Crucial Considerations that have not been considered or researched yet. We have an interesting group of academics and I would take a mixed position of CEO and researcher. 2) Altruism: past, present, propagation: this is a book whose table of contents I already wrote, and would need further research and spelling out each of the 250 sections I have in mind. It is very different in nature from Will's book, or Singer's book. The idea here is not to introduce to EA, but to reason about the history of cooperation and altruism that led to us, and where this can be taken in the future, inclusive by the EA movement. This would be major intellectual undertaking, likely consuming my next three years and doubling as a PHD dissertation. Perhaps, tripling as a series of blog posts, for quick feedback loops and reliable writer motivation. 3) FLI grant proposal: Our proposal intended to increase our understanding psychological theories of human morality in order to facilitate later work in formalizing moral cognition to AIs, a subset of the value loading and control problems of Artificial Generalized Intelligence. We didn't win, so the plan here would be to try to find other funding sources for this research. 4) Accelerate the PHD: For that I need to do 3 field statements, one about the control problem in AI with Stuart, one about altruism with Deacon, and one to be determined, then only the dissertation would be still on the to do list. All these plans scored sufficiently high in my calculations that it is hard to decide between them. Accelerating the PHD has a major disadvantage because it does not increase my funding. The book (via blog posts or not) has a strong advantage in that I think it will have sufficiently new material that it satisfies goal 1 best of all, it is probably the best for the world if

That sounds about right :)

I like your sincerity. The verbosity is something I actually like and quite praised in the human sciences I was raised in, I don't aim for the condensed information writing style. The nascissism I dislike and tried to fix before, but it's hard, it's a mix of a rigid personality trait with a discomfort from having been in the EA movement since long before it was an actual thing, having spent many years giving time resources and attention, and seeing new EAs who don't have knowledge or competence being rewarded (especially financia... (read more)

My experience on Lesswrong indicates that though well intentioned, this would be a terrible policy. The best predictor on Lesswrong if texts of mine would be upvoted or downvoted was wheter someone, in particular username Shminux, would give reasons for their downvote.

There is nothing I dislike or fear more, when I write on Lesswrong than Shminux giving reasons why he's downvoting this time.

Don't get me wrong, write a whole dissertation about what in the content is wrong, or bad, or unformatted, do anything else, but don't say, for instance "Downvoted... (read more)

5
Tom_Ash
Hmm, that's a good point and I don't know what to think any more.

I will write that post once I am finantially secure with some institutional attachment. I think it is too important for me to write while I expect to receive funding as an individual, and don't want people to think "he's saying that because he is not financed by an institution." Also see this.

I think we are falling prey to the transparency fallacy

https://en.wikipedia.org/wiki/Illusion_of_transparency

, the double transparency fallacy,

http://lesswrong.com/lw/ki/double_illusion_of_transparency/

and that there are large inferential gaps in our conversation in both directions.

We could try to close the gaps writing to one another here, but then both of us would end up sometimes taking a defensive stance which could hinder discussion progress. My suggestion is that we do one of these

1) We talk via skype or hangouts to understand each other's mind. 2... (read more)

Do you want to do this soon? You can help us get it done.

0
Evan_Gaensbauer
Yeah, let's do it. If others are rearing to go, I'll help edit or provide feedback or share links or do whatever I can. Between so many major foundations seeming poised to give more money than the rest of the community combined to all our current favored charities across causes, but them being unlikely to directly fund individuals' research, now is an ideal time to plant in the minds of 4-to-6 figure-per-year donors new acceptable targets for their donations.

I'll bite: 1) Transhumanism: The evidence is for the paucity of our knowledge. 2) Status: People are being valued not for the exp value they produce, but by the position they occupy. 3) Analogy: Jargon from Musk, meaning copying and tweaking someone else's idea instead of thinking of a rocket, for instance, from the ground up - follow the chef and cook link. 4) Detonator: Key word was "cling to", they stick with one they had to begin with, demonstrating lack of malleability. 5) Size: The size gives reason to doubt the value of action because to ... (read more)

1
kbog
Transhumanists don't claim that we already know exactly how to improve the human condition, just that we can figure it out. I'm not sure what kind of valuing you're referring to, but it doesn't sound like the same thing as the status seeking you first talked about, which is a question of behavior. Really I don't see where the problem is at all - we don't have any need to "value" people as if we were sacrificing them, sounds like you're just getting worked about who's getting the most praise and adoration! Praise and blame cannot, and never will be, metered out on the basis of pure contribution. They will always be subject to various human considerations, and we'll always have to accept that. There's always a balance to be had between too many and too few original ideas. I see plenty of EAs who try to follow their own ideas instead of contributing to what we already know. Again, can you provide any evidence that this is an actual problem which is playing out in the movement, rather than your own personal opinion? So you expect people to be changing their moral opinions more? Why? Would you complain that civil rights activists cling too much to beliefs in equality? How is this a problem with negative ramifications for the movement, and why does it give you grounds to be "skeptical"? People should be rational, but they shouldn't change their opinions for the sake of changing them. Sometimes they just have good reasons to believe that they are right. I'm sorry but I don't understand how this backs up your point. It seems to actually contradict what you were saying, because if the problems are large then there is a huge cost to ignoring them. I don't even think this makes sense. Opportunistic cost and complexity of the issue are orthogonal issues. No... opportunity cost is a function of expected value, not certainty. I can't tell what it is you would value that forces this dilemma. If you can't figure out how to explain your idea quickly and simply, that should s
4
AGB 🔸
Re. 10), it's worth saying that in the previous post Paul Chistiano and I noted that it wasn't at all obvious to us why you thought individuals were generally cheaper than institutions, with tax treatment versus administrative overhead leading to an unclear and context-specific conclusion. You never replied, so I still don't know why you think this. I would definitely be curious to know why you think this.
0
Diego_Caleiro
9) I ended up replying how to on Lesswrong.

This post is the sort of thing I would expect Crux - the Crucial Considerations Institute we are forming in a few months - to output on a regular basis.

0
Gleb_T
Can you tell me more about Crux? I'm curious about it. My email is gleb@intentionalinsights.org

Continue thinking this is great work and new EAs should always be directed to this :)

Arguing for Cryo as EA seems to be a bottom line reasoning for me.

I can imagine exceptions. For instance: 1) Mr E.A. is an effective altruist who is super productive and gets most enjoyment out of working, and rests by working even more. Expecting to be an emulation with high likelihood, Mr E. A. decided for cryopreservation to give himself a chance of becoming an emulation coalition which would control large fractions of the EM economy, and use these resources for the EA cause on which society has settled after long and careful thought.

2) Rey Cortzvai... (read more)

Thumbs up for this. Creating a labor market for people who are willing to work for causes seems high value to me.

A few years ago, before I spend most, and while Brazil was doing well, I didn't care about money, and as usual I was working and paying my own work out of pocket.

If it was an option then, I would have wanted to work on far future EA, and hedge my bets by asking other people to donate to near future causes on behalf of the work I was doing. I currently lean much more strongly towards far future though, so most of my eggs are in that basket. ... (read more)

Off the top of my mind and without consulting people:

Justin Shovelain, Oliver Habryka, Malcolm Ocean, Roxanne Heston, Miranda Dixon-Luinenburg, Steve Rayhawk, Gustavo Rosa, Stephen Frey, Gustavo Bicalho, Steven Kaas, Bastien Stern, Anne Wissemann, and many others.

If they had not received FLI funding: Kaj Sotala, Katja Grace.

If they needed to transition between institutions/countries: Most of the core EA community.

I have mentioned it as an option for a while, but personally waited for less conflict of interest to actually post about it (at the moment... (read more)

0
Evan_Gaensbauer
If this market can start taking off, encourage each of those persons you mentioned to post an outline of the research they're pursuing on the Effective Altruism Forum or LessWrong or something. Then, potential funders have somewhere to browse and ask questions about the research proposal(s).

Would it be valuable to develop a university level course on AI safety engineering to be implemented in hundreds of universities that use Russell's book worldwide, to attract more talented minds to the field? Which are the steps that would cause this to happen?

Robin Hanson is not mainstream in any sense I can envision, he did take a look at it though :) I asked an economist friend to review, and an economy student reviewed it as well. Check below for the link for the complete google docs if you are an economist who happens to be reading this.

If you are eager to see the other posts in the series, and would like to help them by commenting, feel free to comment in this google docs which contains all the posts in the series.

The posts are already finished, yet, I highly encourage other EAs to create more posts in that document, or suggest changes. I'm not an economist, I was struck by this idea while writing my book on Altruism, and already spent many hours learning more economics to develop it. The goal is to have actual economists carrying this on to distances I cannot.

The question I would ask then is, if you want to influence larger organization, why not governmental organizations, which have the largest quantities of resources that can be flipped by one individual? If you get a technical position in a public policy related organization, you may be responsible for substantial changes in allocation of resources.

1
SydMartin
I think that governmental orgs would be a great way to do this! I do worry that doing this as an individual has it's draw backs. I think getting to this sort of position requires ingraining yourself into a dysfunctional culture and I worry about getting sucked into the dysfunction, or succumbing to the multiple pressures and restraints within such an organization. Whereas an independent organization could remain more objective & focused on effectiveness.

At the end of the day, the metric will always be the same. If you can make the entire red cross more effective, it may be that each unit of your effort was worth it. But if you anticipate more and more donations going to EA recommended charities, then making them even more effective may be more powerful.

See also DavidNash comment.

2
mhpage
Of course. But as I understand it, the hypothesis here is that given (i) the amount of money that will invariably go to sub-optimal charities; and (ii) the likely room for substantial improvements in sub-optimal charities (see DavidNash's comment), that one (arguably) might get more bang for their buck trying to fix sub-optimal charities. I think it's a plausible hypothesis. I'm doubtful that one can make GiveWell charities substantially more effective. Those charities are already using the EA lens. It's the ones that aren't using the EA lens for which big improvements might be made at low cost. EDIT: I suppose I'm assuming that's the OP's hypothesis. I could be wrong.

Except for the purposes of obtaining more epistemic information later on, the general agreement within the EA crowd is that one should invest the vast majority of eggs in one basket, the best basket.

I just want to point out the exact same is the case here, where if someone wants to make a charity more effective, choosing oxfam or the red cross would be a terrible idea, but trying to make AMF, FHI, SCI etc more effective would be a great idea.

Effective altruism is a winners take all kind of thing, where the goal is to make the best better, not to make anyone else be as good as the best.

4
DavidNash
If you can make an organisation that deals with billions of dollars 1% more effective, I think that could have a similar outcome to making an effective charity that works with millions of dollars 1% more effective. There may be more scope for change as well if it isn't that effective to begin with. Also getting higher up an organisation will lead to greater opportunities to change it from within rather than always staying outside because they aren't as efficient.
6
mhpage
This is true with respect to where a rational, EA-inclined person chooses to donate, but I think you're taking it too far here. Even in the best case scenario, there will be MANY people who donate for non-EA reasons. Many of those people will donate to existing, well-known charities such as the Red Cross. If we can make the Red Cross more effective, I can't see how that would not be a net good.

This piece is a simplified version of an academic article Joao Fabiano and I are writing on the future of evolutionary forces, similar in spirit to this one. It will also be the basis of one of the early chapters of my book Altruism: past, present, propagation. We welcome criticism and suggestions of other forces/constraints/conventions that may be operating to interfere or accelerate the long term evolution of coalitions, cooperation, and global altruistic coordination.

1) I see a trend in the way new EAs concerned about the far future think about where to donate money that seems dangerous, it goes:

I am an EA and care about impactfulness and neglectedness -> Existential risk dominates my considerations -> AI is the most important risk -> Donate to MIRI.

The last step frequently involves very little thought, it borders on a cached thought.

How would you be conceiving of donating your X-risk money at the moment if MIRI did not exist? Which other researchers or organizations should be being scrutinized by donors who are X-risk concerned, and AI persuaded?

5
So8res
1) Huh, that hasn't been my experience. We have a number of potential donors who ring us up and ask who in AI alignment needs money the most at the moment. (In fact, last year, we directed a number of donors to FHI, who had much more of a funding gap than MIRI did at that time.) 2) If MIRI disappeared and everything else was held constant, then I'd be pretty concerned about the lack of people focused on the object level problems. (All talk more about why I think this is so important in a little bit, I'm pretty sure at least one other person asks that question more directly.) There'd still be a few people working on the object level problems (Stuart Russell, Stuart Armstrong), but I'd want lots more. In fact, that statement is also true in the actual world! We only have three people on the research team right now, remember, with a fourth joining in August. In other words, if you were to find yourself in a world like this one except without a MIRI, then I would strongly suggest building something like a MIRI :-)

1)Which are the implicit assumptions, within MIRI's research agenda, of things that "currently we have absolutely no idea of how to do that, but we are taking this assumption for the time being, and hoping that in the future either a more practical version of this idea will be feasible, or that this version will be a guiding star for practical implementations"?

I mean things like

  • UDT assumes it's ok for an agent to have a policy ranging over all possible environments and environment histories

  • The notion of agent used by MIRI assumes to some ex

... (read more)
7
So8res
1) The things we have no idea how to do aren't the implicit assumptions in the technical agenda, they're the explicit subject headings: decision theory, logical uncertainty, Vingean reflection, corrigibility, etc :-) We've tried to make it very clear in various papers that we're dealing with very limited toy models that capture only a small part of the problem (see, e.g., basically all of section 6 in the corrigibility paper). Right now, we basically have a bunch of big gaps in our knowledge, and we're trying to make mathematical models that capture at least part of the actual problem -- simplifying assumptions are the norm, not the exception. All I can easily say that common simplifying assumptions include: you have lots of computing power, there is lots of time between actions, you know the action set, you're trying to maximize a given utility function, etc. Assumptions tend to be listed in the paper where the model is described. 2) The FLI folks aren't doing any research; rather, they're administering a grant program. Most FHI folks are focused more on high-level strategic questions (What might the path to AI look like? What methods might be used to mitigate xrisk? etc.) rather than object-level AI alignment research. And remember that they look at a bunch of other X-risks as well, and that they're also thinking about policy interventions and so on. Thus, the comparison can't easily be made. (Eric Drexler's been doing some thinking about the object-level FAI questions recently, but I'll let his latest tech report fill you in on the details there. Stuart Armstrong is doing AI alignment work in the same vein as ours. Owain Evans might also be doing object-level AI alignment work, but he's new there, and I haven't spoken to him recently enough to know.) Insofar as FHI folks would say we're making assumptions, I doubt they'd be pointing to assumptions like "UDT knows the policy set" or "assume we have lots of computing power" (which are obviously simplifying assu

These are very good points, I endorse checking John's comments.

Some additional related points:

1) Joao Fabiano looked recently into acceptance likelihood for papers in the top 5 philosophical journals. It seems that 3-5% is a reasonable range. It is very hard to publish philosophy papers. It seems to be slightly harder to publish in the top philosophy journals than in Nature, Behavioral and Brain Sciences, or Science magazine, and this is after the filter of 6 positions available for 300 candidates that selects PHD candidates in philosophy (harder than Harvard medicine or economics).

2) Bostrom arguably became very ... (read more)

1
zackrobinson
"1) Joao Fabiano looked recently into acceptance likelihood for papers in the top 5 philosophical journals. It seems that 3-5% is a reasonable range. It is very hard to publish philosophy papers. It seems to be slightly harder to publish in the top philosophy journals than in Nature, Behavioral and Brain Sciences, or Science magazine, and this is after the filter of 6 positions available for 300 candidates that selects PHD candidates in philosophy (harder than Harvard medicine or economics)." This is a very real problem. Many people outside of philosophy do not realize how difficult it is to become an actual tenured philosopher, and even if that happens, to become a regularly-published philosopher. Philosophy itself is already highly self-selecting (highest average GRE of any grad school-bound college students), and the acceptance rates are very, very low. Further, only about half of those students who are accepted to the top PhD programs complete them. Of those who do complete them, I'd say about half (of the top programs) end up with tenure-track positions. Those who do get such positions may or may not have difficulty gaining publication, but it isn't a certainty that their ideas will be widespread in any real way. So, for Joe Smart sitting at home on his couch, wanting his great ideas to be read by a lot of smart people, becoming a philosopher is probably a terrible way of accomplishing his goal. "4) For people who consider themselves intellectual potentials and intend to continue in academia, my suggestion is to create a table of contents for a book, and instead of going ahead and writing the chapters, find the closest equivalent of some chapter that could become a paper, and try to write a paper about that. If you get accepted, this develops your career, and allows you to be one of the stand-outs like Ord, MacAskill and Bostrom who will end up working in the top universities. If you continue to be systematically rejected, you can still get around by publishi

More important than my field not being philosophy anymore (though I have two degrees in philosophy, and identify as a philosopher) the question you could have asked there is why would you want a philosophical audience to begin with? Seems to me there is more low hanging fruits in nearly any other area in terms of people who could become EAs. Philosophers have an easier time doing that, but attracting the top people in econ, literature, visual arts and others who may enjoy reading the occasional public science books is much less replaceable.

I've left the field of philosophy (where I was mostly so I could research what seemed interesting and not what the university wanted, as Chalmers puts it "studying the philosophy of x" where x is what interests me at any time) and am now in biological anthropology. It seems that being a professor in non-philosophy fields is much easier than in philosophy, from my many years researching the topic. Also switching fields between undergrad and grad school is easy, in case someone reading this does not know.

0
RyanCarey
Interesting. I'm sure you could carve out an interesting niche in that area. One immediately obvious issue in that area is how modern humans use various environmental resources. More distantly relevant issues that you might still be closer to than any other current EA academics would be different examples of the culture of science, or innovation, and human views on our place in relation to evolution, including transhumanism. I'm sure there are others.

Biological Anthropology, with an adviser whose latest book is in philosophy of mind, the next book on information theory, the previous book on - of all things - biological anthropology, and most of his career was as a semioticist and neuroscientist. My previous adviser was a physicist in the philosophy of physics who turned into a philosopher of mind. My main sources of inspiration are Bostrom and Russell, who defy field borders. So I'm basically studying whatever you convince me makes sense in the intersection of interestingly complex and useful for the world. Except for math, code and decision theory, which are not my comparative advantage, specially not among EAs.

0
Brian_Tomasik
Thanks for asking for suggestions. :) I would tend to focus on AGI-related topics, though you may have specific alternate ideas that are compelling for reasons that I can't see from a distance. In addition to AGI safety, studying political dynamics of AGI takeoff (including de novo AGI, emulations, etc.) could be valuable. I suggested a few very general AGI research topics here and here. Some broader though perhaps less important topics are here.

What are the unsolved problems related to infinite ethics that might be worth tackling as an academic? Some relevant writings on this topic to see what the field looks like

0
RyanCarey
I would ask Amanda MacAskill. On the topic of interesting questions about large utilities, I think could conceivably be useful to analyse the notion that it might be good to refuse one 'Pascal's Mugging' proposition in order to pursue other ones, and to see what this might imply, as well as to analyse how mugging works out in game theory or decsion theory to decide whether its adversarial nature is important.

I am not considering what Bostrom/Grace/Besinger/ do philosophy strictu sensu in this question.

After repleaceability considerations have been used at Ben Todd and Will Mac Askill's theses at Oxford, and Nick Beckstead made the philosophical case for the far future, is there still large marginal return to be had on doing research on something that is philosophy strictu sensu?

I ask this because my impression is that after Parfit, Singer, Unger, Ord, Mac Askill and Todd we have run out of efforts that have great consequential impacts in philosophical discour... (read more)

0
RyanCarey
Personally, I wouldn't feel bad if we left technical philosophy to MacAskill and Ord for a while, as they're surely going to keep doing it. But maybe you want to get professorship somehow, and if so, then your choices are reduced.
0
zdgroff
It seems like you may have more insight than anyone else on whether you should go into philosophy. If you have a high-impact idea or set of ideas that you think you can contribute, perhaps you can go in. Before MacAskill and Ord, I don't think many people thought there was an applicable and useful argument to be made in philosophy, but they proved that wrong.
Load more