Crossposted from opentheory.net
STV is Qualia Research Institute‘s candidate for a universal theory of valence, first proposed in Principia Qualia (2016). The following is a brief discussion of why existing theories are unsatisfying, what STV says, and key milestones so far.
I. Suffering is a puzzle
We know suffering when we feel it — but what is it? What would a satisfying answer for this even look like?
The psychological default model of suffering is “suffering is caused by not getting what you want.” This is the model that evolution has primed us toward. Empirically, it appears false (1)(2).
The Buddhist critique suggests that most suffering actually comes from holding this as our model of suffering. My co-founder Romeo Stevens suggests that we create a huge amount of unpleasantness by identifying with the sensations we want and making a commitment to ‘dukkha’ ourselves until we get them. When this fails to produce happiness, we take our failure as evidence we simply need to be more skillful in controlling our sensations, to work harder to get what we want, to suffer more until we reach our goal — whereas in reality there is no reasonable way we can force our sensations to be “stable, controllable, and satisfying” all the time. As Romeo puts it, “The mind is like a child that thinks that if it just finds the right flavor of cake it can live off of it with no stomach aches or other negative results.”
Buddhism itself is a brilliant internal psychology of suffering (1)(2), but has strict limits: it’s dogmatically silent on the influence of external factors on suffering, such as health, relationships, or anything having to do with the brain.
The Aristotelian model of suffering & well-being identifies a set of baseline conditions and virtues for human happiness, with suffering being due to deviations from these conditions. Modern psychology and psychiatry are tacitly built on this model, with one popular version being Seligman’s PERMA Model: P – Positive Emotion; E – Engagement; R – Relationships; M – Meaning; A – Accomplishments. Chris Kresser and other ‘holistic medicine’ practitioners are synthesizing what I would call ‘Paleo Psychology’, which suggests that we should look at our evolutionary history to understand the conditions for human happiness, with a special focus on nutrition, connection, sleep, and stress.
I have a deep affection for these ways of thinking and find them uncannily effective at debugging hedonic problems. But they’re not proper theories of mind, and say little about the underlying metaphysics or variation of internal experience.
Neurophysiological models of suffering try to dig into the computational utility and underlying biology of suffering. Bright spots include Friston & Seth, Panksepp, Joffily, and Eldar talking about emotional states being normative markers of momentum (i.e. whether you should keep doing what you’re doing, or switch things up), and Wager, Tracey, Kucyi, Osteen, and others discussing neural correlates of pain. These approaches are clearly important parts of the story, but tend to be descriptive rather than predictive, either focusing on ‘correlation collecting’ or telling a story without grounding that story in mechanism.
QRI thinks not having a good answer to the question of suffering is a core bottleneck for neuroscience, drug development, and next-generation mental health treatments, as well as philosophical questions about the future direction of civilization. We think this question is also much more tractable than people realize, that there are trillion-dollar bills on the sidewalk, waiting to be picked up if we just actually try.
II. QRI’s model of suffering – history & roadmap
What does “actually trying” to solve suffering look like? I can share what we’ve done, what we’re doing, and our future directions.
QRI.2016: We released the world’s first crisp formalism for pain and pleasure: the Symmetry Theory of Valence (STV)
QRI had a long exploratory gestation period as we explored various existing answers and identified their inadequacies. Things started to ‘gel’ as we identified and collected core research lineages that any fundamentally satisfying answer mustengage with.
A key piece of the puzzle for me was Integrated Information Theory (IIT), the first attempt at a formal bridge between phenomenology and causal emergence (Tononi et. al 2004, 2008, 2012). The goal of IIT is to create a mathematical object ‘isomorphic to’ a system’s phenomenology — that is to say, to create a perfect mathematical representation of what it feels like to be something. If it’s possible to create such a mathematical representation of an experience, then how pleasant or unpleasant the experience is should be ‘baked into’ this representation somehow.
In 2016 I introduced the Symmetry Theory of Valence (STV) built on the expectation that, although the details of IIT may not yet be correct, it has the correct goal — to create a mathematical formalism for consciousness. STV proposes that, given such a mathematical representation of an experience, the symmetry of this representation will encode how pleasant the experience is (Johnson 2016). STV is a formal, causal expression of the sentiment that “suffering is lack of harmony in the mind” and allowed us to make philosophically clear assertions such as:
- X causes suffering because it creates dissonance, resistance, turbulence in the brain/mind.
- If there is dissonance in the brain, there is suffering; if there is suffering, there is dissonance in the brain. Always.
This also let us begin to pose first-principles, conceptual-level models for affective mechanics: e.g., ‘pleasure centers’ function as pleasure centers insofar as they act as tuning knobs for harmony in the brain.
QRI.2017: We figured out how to apply our formalism to brains in an elegant way: CDNS
We had a formal hypothesis that harmony in the brain feels good, and dissonance feels bad. But how do we measure harmony and dissonance, given how noisy most forms of neuroimaging are?
An external researcher, Selen Atasoy, had the insight to use resonance as a proxy for characteristic activity. Neural activity may often look random— a confusing cacophony— but if we look at activity as the sum of all natural resonances of a system we can say a great deal about how the system works, and which configuration the system is currently in, with a few simple equations. Atasoy’s contribution here was connectome-specific harmonic waves (CSHW), an experimental method for doing this with fMRI (Atasoy et. al 2016; 2017a; 2017b). This is similar to how mashing keys on a piano might produce a confusing mix of sounds, but through applying harmonic decomposition to this sound we can calculate which notes must have been played to produce it. There are many ways to decompose brain activity into various parameters or dimensions; CSHW’s strength is it grounds these dimensions in physical mechanism: resonance within the connectome. (See also work by Helmholtz, Tesla, and Lehar.)
QRI built our ‘Consonance Dissonance Noise Signature’ (CDNS) method around combining STV with Atasoy’s work: my co-founder Andrés Gomez Emilsson had the key insight that if Atasoy’s method can give us a power-weighted list of harmonics in the brain, we can take this list and do a pairwise ‘CDNS’ analysis between harmonics and sum the result to figure out how much total consonance, dissonance, and noise a brain has (Gomez Emilsson 2017). Consonance is roughly equivalent to symmetry (invariance under transforms) in the time domain, and so the consonance between these harmonics should be a reasonable measure for the ‘symmetry’ of STV. This process offers a clean, empirical measure for how much harmony (and lack thereof) there is in a mind, structured in a way that lets us be largely agnostic about the precise physical substrate of consciousness.
With this, we had a full empirical theory of suffering.
QRI.2018: We invested in the CSHW paradigm and built ‘trading material’ for collaborations
We had our theory, and tried to get the data to test it. We decided that if STV is right, it should let us build better theory, and this should open doors for collaboration. This led us through a detailed exploration of the implications of CSHW (Johnson 2018a), and original work on the neuroscience of meditation (Johnson 2018b) and the phenomenology of time (Gomez Emilsson 2018).
QRI.2019: We synthesized a new neuroscience paradigm (Neural Annealing)
2019 marked a watershed for us in a number of ways. On the theory side, we realized there are many approaches to doing systems neuroscience, but only a few really good ones. We decided the best neuroscience research lineages were using various flavors of self-organizing systems theory to explain complex phenomena with very simple assumptions. Moreover, there were particularly elegant theories from Atasoy, Carhart-Harris, and Friston, all doing very similar things, just on different levels (physical, computational, energetic). So we combined these theories together into Neural Annealing (Johnson 2019), a unified theory of music, meditation, psychedelics, trauma, and emotional updating:
Annealing involves heating a metal above its recrystallization temperature, keeping it there for long enough for the microstructure of the metal to reach equilibrium, then slowly cooling it down, letting new patterns crystallize. This releases the internal stresses of the material, and is often used to restore ductility (plasticity and toughness) on metals that have been ‘cold-worked’ and have become very hard and brittle— in a sense, annealing is a ‘reset switch’ which allows metals to go back to a more pristine, natural state after being bent or stressed. I suspect this is a useful metaphor for brains, in that they can become hard and brittle over time with a build-up of internal stresses, and these stresses can be released by periodically entering high-energy states where a more natural neural microstructure can reemerge.
This synthesis allowed us to start discussing not only which brain states are pleasant, but what processes are healing.
QRI.2020: We raised money, built out a full neuroimaging stack, and expanded the organization
In 2020 the QRI technical analysis pipeline became real, and we became one of the few neuroscience groups in the world able to carry out a full CSHW analysis in-house, thanks in particular to hard work by Quintin Frerichs and Patrick Taylor. This has led to partnerships with King’s College London, Imperial College London, National Institute of Mental Health of the Czech Republic, Emergent Phenomenology Research Consortium, as well as many things in the pipeline. 2020 and early 2021 also saw us onboard some fantastic talent and advisors.
III. What’s next?
We’re actively working on improving STV in three areas:
- Finding a precise physical formalism for consciousness. Asserting that symmetry in the mathematical representation of an experience corresponds with the valence of the experience involves a huge leap in clarity over other theories. But we also need to be able to formally generate this mathematical representation. I’ve argued previously against functionalism and for a physicalist approach to consciousness (partially echoing Aaronson), and Barrett, Tegmark, and McFadden offer notable arguments suggesting the electromagnetic field may be the physical seat of consciousness because it’s the only field that can support sufficient complexity. We believe determining a physical formalism for consciousness is intimately tied to the binding problem, and have conjectures I’m excited to test.
- Building better neuroscience proxies for STV. We’ve built our empirical predictions around the expectation that consonance within a brain’s connectome-specific harmonic waves (CSHW) will be a good proxy for the symmetry of that mind’s formal mathematical representation. We think this is a best-in-the-world compression for valence. But CSHW rests on a chain of inferences about neuroimaging and brain structure, and using it to discuss consciousness rests on further inferences still. We think there’s room for improvement.
- Building neurotech that can help people. The team may be getting tired of hearing me say this, but: better philosophy should lead to better neuroscience, and better neuroscience should lead to better neurotech. STV gives us a rich set of threads to follow for clear neurofeedback targets, which should allow for much more effective closed-loop systems, and I am personally extraordinarily excited about the creation of technologies that allow people to “update toward wholesome”, with the neuroscience of meditation as a model.
The Symmetry Theory of Valence sounds wrong to me and is not substantiated by any empirical research I am aware of. (Edited to be nicer.) I'm sorry to post a comment so negative and non-constructive, but I just don't want EA people to read this and think it is something worth spending time on.
Credentials: I'm doing a PhD in Neuroscience and Psychology at Princeton with a focus on fMRI research, I have a masters in Neuroscience from Oxford, I've presented my fMRI research projects at multiple academic conferences, and I published a peer reviewed fMRI paper in a mainstream journal. As far as I can tell, nobody at the Qualia Research Institute has a PhD in Neuroscience or has industry experience doing equivalent level work. Keeping in mind credentialism is bad, I am still pointing out their lack of neuroscience credentials compared to mine because I am confused by how overwhelmingly confident they are in their claims, their incomprehensible use of neuro jargon, and how dismissive they are of my expertise. (Edited to be nicer.) https://www.qualiaresearchinstitute.org/team
There are a lot of things I don't understand about STV, but the primary one is:
- If there is dissonance in the
... (read more)Hi, all. Talk is cheap, and EA Forum karma may be insufficiently nuanced to convey substantive disagreements.
I've taken the liberty to sketch out several forecasting questions that might reflect underlying differences in opinion. Interested parties may wish to forecast on them (which the EA Forum should allow you to do directly, at least on desktop) and then make bets accordingly.
Feel free to also counterpropose (and make!) other questions if you think the existing question operationalizations are not sufficient (I'm far from knowledgeable in this field!).
Hi Linch, that’s very well put. I would also add a third possibility (c), which is “is STV false but generative.” — I explore this a little here, with the core thesis summarized in this graphic:
I.e., STV could be false in a metaphysical sense, but insofar as the brain is a harmonic computer (a strong reframe of CSHW), it could be performing harmonic gradient descent. Fully expanded, there would be four cases:
STV true, STHR true
STV true, STHR false
STV false, STHR true
STV false, STHR false
Of course, ‘true and false’ are easier to navigate if we can speak of absolutes; STHR is a model, and ‘all models are wrong; some are useful.’
For what it's worth, I read this comment as constructive rather than non-constructive.
If I write a long report and an expert in the field think that the entire premise is flawed for specific technical reasons, I'd much rather them point this out rather than for them to worry about niceness and then never getting around to mentioning it, thus causing my report to languish in obscurity without me knowing why (or worse, for my false research to actually be used!)
I’m a bit hesitant to upvote this comment given how critical it is [was] + how little I know about the field (and thus whether the criticism is deserved), but I’m a bit relieved/interested to see I wasn’t the only one who thought it sounded really confusing/weird. I have somewhat skeptical priors towards big theories of consciousness and suffering (sort of/it’s complicated) + towards theories that rely on lots of complicated methods/jargon/theory (again, sort of/with caveats)—but I also know very little about this field and so I couldn’t really judge. Thus, I’m definitely interested to see the opinions of people with some experience in the field.
Haha, I certainly wouldn't label what you described/presented as "timecube weird." To be honest, I don't have a very clear cut set of criteria, and upon reflection it's probable that the prior is a bit over-influenced by my experiences with some social science research and theory as opposed to hard science research/theory. Additionally, it's not simply that I'm skeptical of whether the conclusion is true, but more generally my skepticism heuristics for research is about whether whatever is being presented is "A) novel/in contrast with existing theories or intuitions; B) is true; and/or C) is useful." For example, some theory might be basically rehashing what existing research already has come to consensus on but simply worded in a very different way that adds little to existing research (aside from complexity); alternatively, something could just be flat out wrong; alternatively, something could be technically true and novel as explicitly written, but that is not very useful (e.g., tautological definitions), whereas the common interpretation is wrong (but would be useful if it were right).
Still, two of the key features here that contributed to my mental yellow flags were:
- The emphas
... (read more)People are asking for object-level justifications for the Symmetry Theory of Valence:
The first thing to mention is that the Symmetry Theory of Valence (STV) is *really easy to strawman*. It really is the case that there are many near enemies of STV that sound exactly like what a naïve researcher who is missing developmental stages (e.g. is a naïve realist about perception) would say. That we like pretty symmetrical shapes of course does not mean that symmetry is at the root of valence; that we enjoy symphonic music does not mean harmony is "inherently pleasant"; that we enjoy nice repeating patterns of tactile stimulation does not mean, well, you get the idea...
The truth of course is that at QRI we really are meta-contrarian intellectual hipsters. So the weird and often dumb-sounding things we say are already taking into account the criticisms people in our people-cluster would make and are taking the conversation one step further. For instance, we think digital computers cannot be conscious, but this belief comes from entirely different arguments than those that justify such beliefs out there. We think that the "energy body" is real and important, except that we in... (read more)
Edit: This comment now makes less sense, given that Abby has revised the language of her comment.
Abby,
I strongly endorse what you say in your last paragraph:
However, I'd like to push back on the tone of your reply. If you're sorry for posting a negative non-constructive comment, why not try to be a bit more constructive? Why not say something like "I am deeply skeptical of this theory and do not at this moment think it's worth EAs spending time on. [insert reasons]. I would be willing to change my view if there was evidence."
Apologies for being pedantic, but I think it's worth the effort to try and keep the conversation on the forum as constructive as possible!
Hi Jpmos,
I think context is important here. This is not an earnest but misguided post from an undergrad with big ideas and little experience. This is a post from an organization trying to raise hundreds of thousands of dollars. You can check out their website if you want, the front page has a fundraising advertisement.
Further, there are a lot of fancy buzzwords in this post ("connectome!") and enough jargon that people unfamiliar with the topic might think there is substance here that they just don't understand (see Harrison's comment: "I also know very little about this field and so I couldn't really judge").
As somebody who knows a lot about this field, I think it's important that my opinion on these ideas is clearly stated. So I will state it again.
There is no a priori reason to believe any of the claims of STV. There is no empirical evidence to support STV. To an expert, these claims do not sound "interesting and plausible but unproven", they sound "nonsensical and presented with baffling confidence".
People have been observing brain oscillations at different frequencies and at different powers for about 100 years. These oscillations have been associated w... (read more)
This reads to me as insinuating fraud, without much supporting evidence.
I appreciate that in other comments you followed up with more concrete criticisms, but this still feels against the "Keep EA Weird" spirit to me. If we never spend a million or two on something that turns out to be nonsense, we wouldn't have applied hits-based giving very well.
(Despite the username, I have no affiliation with QRI. I'll admit to finding the problem worth working on. )
Keeping EA honest and rigorous is much higher priority. Making excuses for incompetence or lack of evidence base is the opposite of EA.
I agree that honesty is more important than weirdness. Maybe I’m being taken, but I see miscommunication and not dishonesty from QRI.
I am not sure what an appropriate standard of rigor is for a preparadigmatic area. I would welcome more qualifiers and softer claims.
Hi all, I messaged some with Holly a bit about this, and what she shared was very helpful. I think a core part of what happened was a mismatch of expectations: I originally wrote this content for my blog and QRI’s website, and the tone and terminology was geared toward “home team content”, not “away team content”. Some people found both the confidence and somewhat dense terminology offputting, and I think that’s reasonable of them to raise questions. As a takeaway, I‘ve updated that crossposting involves some pitfalls and intend to do things differently next time.
Hi Mike,
I am comfortable calling myself "somebody who knows a lot about this field", especially in relation to the average EA Forum reader, our current context.
I respect Karl Friston as well, I'm looking forward to reading his thoughts on your theory. Is there anything you can share?
The CSHW stuff looks potentially cool, but it's separate from your original theory, so I don't want to get too deep into it here. The only thing I would say is that I don't understand why the claims of your original theory cannot be investigated using standard (cheap) EEG techniques. This is important if a major barrier to finding empirical evidence for your theory is funding. Could you explain why standard EEG is insufficient to investigate the synchrony of neuronal firing during suffering?
I was very aggressive with my criticism of your theory, partially because I think it is wrong (again, the basis of your theory, "the symmetry of this representation will encode how pleasant the experience is", makes no sense to me), but also because of how confidently you describe your theory with no empirical evidence. So I happily accept being called arrogant and would also happily accept being sh... (read more)
Hi Mike! I appreciate your openness to discussion even though I disagree with you.
Some questions:
1. The most important question: Why would synchrony between different brain areas involved in totally different functions be associated with subjective wellbeing? I fundamentally don't understand this. For example, asynchrony has been found to be useful in memory as a way of differentiating similar but different memories during encoding/rehearsal/retrieval. It doesn't seem like a bad thing that the brain has a reason to reduce, the way it has reasons to reduce prediction errors. Please link to brain studies that have found asynchrony leads to suffering.
2. If your theory is focused on neural oscillations, why don't you use EEG to measure the correlation between neural synchrony and subjective experience? Surely EEG is a more accurate method and vastly cheaper than fMRI?
3. If you are funding constrained, why are none of your collaborators willing to run this experiment for you? Running fMRI and EEG experiments at Princeton is free. I see you have multiple Princeton affiliates on your team, and we even have Michael Graziano as a faculty member who is deeply in... (read more)
This sounds overwhelmingly confident to me, especially since you have no evidence to support either of these claims.
Hi Mike, I really enjoy your and Andrés's work, including STV, and I have to say I'm disappointed by how the ideas are presented here, and entirely unsurprised at the reaction they've elicited.
There's a world of a difference between saying "nobody knows what valence is made out of, so we're trying to see if we can find correlations with symmetries in imaging data" (weird but fascinating) and "There is an identity relationship between suffering and disharmony" (time cube). I know you're not time cube man, because I've read lots of other QRI output over the years, but most folks here will lack that context. This topic is fringe enough that I'd expect everything to be extra-delicately phrased and very well seasoned with ifs and buts.
Again, I'm a big fan of QRI's mission, but I'd be worried about donating I if I got the sense that the organization viewed STV not as something to test, but as something to prove. Statistically speaking, it's not likely that STV will turn out to be the correct mechanistic grand theory of valence, simply because it's the first one (of hopefully many to come). I would like to know:
- When do you expect to be able to share the first set of empirical results, and
... (read more)Andrés’s STV presentation to Imperial College London’s psychedelics research group is probably the best public resource I can point to on this right now. I can say after these interactions it’s much more clear that people hearing these claims are less interested in the detailed structure of the philosophical argument, and more in the evidence, and in a certain form of evidence. I think this is very reasonable and it’s something we’re finally in a position to work on directly: we spent the last ~year building the technical capacity to do the sorts of studies we believe will either falsify or directly support STV.
I read this post and the comments that have followed it with great interest.
I have two major, and one minor, worries about QRI's research agenda I hope you can clarify. First, I am not sure exactly which question you are trying to answer. Second, it's not clear to me why you think this project is (especially) important. Third, I can't understand what STV is about because there is so much (undefined) technical jargon.
1. Which question is QRI trying to answer?
You open by saying:
This makes me think you want to identify what suffering is, that is, what it consists in. But you then immediately raise Buddhist and Arisotlean theories of what causes suffering - a wholly different issue. FWIW, I don't see anything deeply problematic in identifying what suffering, and related terms, refer to. Valence just refers to how good/bad you feel (the intrinsic pleasurableness/displeasurableness of your experience); happiness is feeling overall good; suffering is feeling overall bad. I don't find anything dissatisfying about these. Valence refers to something subjective. ... (read more)
Hi Michael, I appreciate the kind effortpost, as per usual. I’ll do my best to answer.
Our intent is to define phenomenological valence in a fully formal way, with the template being physics: we wish to develop our models such that we can speak of pain and pleasure with all the clarity, precision, and rigor as we currently describe photons and quarks and fields.
This may sound odd, but physics is a grand success story of formalization, and we essentially wish to apply the things that worked in physics, to phenomenology. Importantly, physics has a strong tradition of using symmetry considerations to inform theory. STV borrows squarely from this tradition (see e.g. my write up on Emmy Noether).
Valence is subjective as you note, but that doesn’t mean it’s arbitrary;... (read more)
I feel like it's important to highlight two things QRI people have said. These statements illustrate why STV sounds extremely implausible to me.
"STV makes a big jump in that it assumes the symmetry of this mathematical object corresponds to how pleasurable the experience it represents is. This is a huge, huge, huge jump, and cannot be arrived at by deduction; none of my premeses force this conclusion. We can call it an educated guess. But, it is my best educated guess after thinking about this topic for about 7 years before posting my theory. I can say I’m fully confident the problem is super important and I’m optimistic this guess is correct, for many reasons, but many of these reasons are difficult to put into words."
"I started out very skeptical of STV myself, and in fact it took about three years of thinking it through in light of many meditation and exotic high-energy experiences to be viscerally convinced that it's pointing in the right direction."
These are not satisfying arguments.
Hi Abby,
I feel we’ve been in some sense talking past each other from the start. I think I bear some of the responsibility for that, based on how my post was written (originally for my blog, and more as a summary than an explanation).
I’m sorry for your frustration. I can only say I’m not intentionally trying to frustrate you, but that we appear to have very different styles of thinking and writing and this may have caused some friction, and I have been answering object-level questions from the community as best I can.
Hi Abby, thanks for the questions. I have direct answers to 2,3,4, and indirect answers to 1 and 5.
1a. Speaking of the general case, we expect network control theory to be a useful frame for approaching questions of why certain sorts of activity in certain regions of the brain are particularly relevant for valence. (A simple story: hedonic centers of the brain act as ‘tuning knobs’ toward or away from global harmony. This would imply they don’t intrinsically create pleasure and suffering, merely facilitate these states.) This paper from the Bassett lab is the best intro I know of to this;
1b. Speaking again of the general case, asynchronous firing isn’t exactly identical to the sort of dissonance we’d identify as giving rise to suffering: asynchronous firing could be framed as in uncorrelated firing, or ‘non-interacting frequency regimes’. There’s a really cool paper asserting that the golden mean is the optimal frequency ratio for non-interaction, and some applications to EEG work, in case you’re curious. What we’re more interested in is frequency combinations that are highly interacting, and lacking a common basis set. An example would be playing the C and C# keys on a piano. This... (read more)
Hi Mike,
Thanks again for your openness to discussion, I do appreciate you taking the time. Your responses here are much more satisfying and comprehensible than your previous statements, it's a bit of a shame we can't reset the conversation.
1a. I am interpreting this as you saying there are certain brain areas that, when activated, are more likely to result in the experience of suffering or pleasure. This is the sort of thing that is plausible and possible to test.
1b. I think you are making a mistake by thinking of the brain like a musical instrument, and I really don't like how you're assuming discordant brain oscillations "feel bad" the way discordant chords "sound bad". (Because as I've stated earlier, there's no evidence linking suffering to dissonance, and as you've stated previously, you made a massive jump in reasoning here.) But this is the clearest you have explained your thinking on this question so far, which I do appreciate.
1c. I am confused here. I did not ask whether dissonance in VWFA causes dissonance in FFA. I asked how dissonance between the two regions causes suffering. What does it mean neurologically to have dissonance within a specific brain ... (read more)
This makes absolutely no sense on its face. I am not a neuroscience expert. I am not a consciousness expert. I do not need to be to say that these conclusions just do not follow.
To recap what you said: You start by saying that, if you could make a complete mathematical representation of the brain (IIT), it would be symmetric to the physical manifestation of the brain, and therefore pleasure would be included in the... (read more)
Hi Holly, I’d say the format of my argument there would be enumeration of claims, not e.g. trying to create a syllogism. I’ll try to expand and restate those claims here:
A very important piece of this is assuming there exists a formal structure (formalism) to consciousness. If this is true, STV becomes a lot more probable. If it isn’t, STV can’t be the case.
Integrated Information Theory (IIT) is the most famous framework for determining the formal structure to an experience. It does so by looking at the causal relationships between components of a system; the more a system’s parts demonstrate ‘integration’ (which is a technical, mathematical term that tries to define how much a system’s parts interact with its other parts), the more conscious the system is.
I didn’t make IIT, I don’t know if it’s true, and I actually suspect it might not be true (I devoted a section of Principia Qualia to explaining IIT, and another section to critiques of IIT). But it’s a great example of an attempt to formalize phenomenology, and I think the project or overall frame of IIT (the idea of consciousness being the sort of thing that one can apply formal mathematics to) is correct even if its implementa... (read more)
I feel like your explanations are skipping a bunch of steps that would help folks understand where you're coming from. FWIW, here's how I make sense of STV:
- Neuroscience can tell us that some neurons light up when we eat chocolate, but it doesn't tell us what it is about the delicious experience of chocolate that makes it so wonderful. "This is what sugar looks like" and "this is the location of the reward center" are great descriptions of parts of the process, but they don't explain why certain patterns of neural activations feel a certain way.
- Everyone agrees that clearly, certain activation patterns do feel a certain way. Quite plausibly, this isn't just a brain thing but more fundamental, and evolution simply recruited the relationship to build RL agents like ourselves. And yet, almost nobody has tried to figure out how exactly the patterns relate to the experiences. Of course, that's because we struggle with both sides of the equation: on the neural side of the equals-sign, the data is incomplete and noisy; on the experience side, what do we even put to represent "delicious"?
- But we have to start somewhere, so we simplify. On the neural side, we look for the simplest kind of sign
... (read more)This is a message I received in private conversation by someone who I trust reasonably highly in terms of general epistemics. I'm reposting it here because it goes against the general "vibe" of the EAF and it's good to get well-informed contrarian opinions.
... (read more)Just a quick comment in terms of comment flow: there’s been a large amount of editing of the top comment, and some of the replies that have been posted may not seem to follow the logic of the comment they‘re attached to. If there are edits to a comment that you wish me to address, I’d be glad if you made a new comment. (If you don’t, I don’t fault you but I may not address the edit.)
Disclaimer: I'm not very familiar with either QRI's research or neuroscience, but in the spirit of Cunningham's Law:
QRI's research seems to predicated on the idea that moral realism and hedonistic utilitarianism are true. I'm very skeptical about both of these, and I think QRI's time would be better spent working on the question of whether these starting assumptions are true in the first place.
Hi Samuel,
I’d say there’s at least some diversity of views on these topics within QRI. When I introduced STV in PQ, I very intentionally did not frame it as a moral hypothesis. If we’re doing research, best to keep the descriptive and the normative as separate as possible. If STV is true it may make certain normative frames easier to formulate, but STV itself is not a theory of morality or ethics.
One way to put this is that when I wear my philosopher’s hat, I’m most concerned about understanding what the ‘natural kinds’ (in Plato’s terms) of qualia are. If valence is a natural kind (similar to how a photon or electromagnetism are natural kinds), that’s important knowledge about the structure of reality. My sense is that ‘understanding what reality’s natural kinds are’ is prior to ethics: first figure out what is real, and then everything else (such as ethics and metaethics) becomes easier.
In terms of specific ethical frames, we do count among QRI some deeply committed hedonistic utilitarians. I see deep value in that frame although I would categorize myself as closer to a virtue ethicist.