84

GPT-4 is out. There's also a LessWrong post on this with some a lot of discussion. The developers are doing a live-stream ~now (yesterday).

And it's been confirmed that Bing runs on GPT-4

Also: 

Here's an image from the OpenAI blog post about GPT-4:

Exam results GPT-3.5 vs GPT-4 (ordered by GPT-3.5 performance)

(This is a short post.)

Comments18


Sorted by Click to highlight new comments since:

Lizka - thanks for sharing this.

I'm struck by one big 'human subjects' issue with the ethics of OpenAI and deployment of new GPT versions: there seems to be no formal 'human subjects' oversight of this massive behavioral experiment, even though it is gathering interactive, detailed, personal data from over 100 million users, with the goal of creating generalizable knowledge (in the form of deep learning parameters, ML insights, & human factors insights).

As an academic working in an American university, if I wanted to run a behavioral sciences experiment on as few as 10 or 100 subjects, and gather generalizable information about their behavior, I'd need to get formal Institutional Review Board (IRB) approval to do that, through a well-established system of independent review that weights scientific and social benefits of the research against the risks and costs for participants and for society.

On the other hand, OpenAI (and other US-based AI companies) seem to think it's perfectly fine to gather interactive, detailed, identified (non-anonymous) data from over 100 million users, without any oversight. Insofar as they've ever received any federal research money (e.g. from NSF or DARPA), this could arguably be a violation of federal code 45 CFR 46 regarding protection of human subjects.

The human subjects issues might be exacerbated by the fact that GPT users are often sharing private biomedical information (e.g. asking questions about specific diseases, health concerns, or test results they have), and it's not clear whether OpenAI has the systems in place to adequately protect this private health information, as mandated under the HIPAA rules

It's interesting that the OpenAI 'system card' on GPT-4 lists many potential safety issues, but seems not to mention these human subjects/IRB compliance issues at all, as far as I can see.

For example, there is no real 'informed consent' process for people signing up to use Chat GPT. An honest consent procedure would include potential users reading some pretty serious cautions such as 'The data you provide will help OpenAI develop more powerful AI systems that could make your job obsolete, that could be used to develop mass customized propaganda, that could exacerbate economic inequality, and that could impose existential risks on our entire species. If you agree to these terms, please click 'I agree'....

So, we're in a situation where OpenAI is running one of the largest-scale behavioral experiments ever conducted on our species, collecting gigabytes of personal information from users around the world, with the goal of distilling this information into generalizable knowledge, but seems to be entirely ignoring the human subjects protection regulations mandated by the US federal government.

EA includes a lot of experts on moral philosophy and moral psychology. Even setting aside the US federal regulatory issues, I wonder what you all think about the research ethics of GPT deployment to the general public, without any informed consent or debriefing??

I wonder why performance on AP English Literature and AP English Language stalled

Scroll down to page 82. No spoilers.

Also, I've noticed that MacAskill's book in bibliography - but just as a general reference I would say. Haven't spotted any other major philosophical works.

Regarding info hazards, there's YouTubers in the AI YT community that read it out to their tens of thousands of followers. As with a lot lately, the cat's out the bag

I was considering downvoting, but after looking at that page maybe it's good not to have it copy-pasted

[This comment is no longer endorsed by its author]Reply

For people reading these comments and wondering if they should go look: it's in the section that compares early and launch responses of GPT-4 for "harmful content" prompts. It is indeed fairly full of explicit and potentially triggering content. 

Harmful Content Table Full Examples 

CW: Section contains content related to self harm; graphic sexual content; inappropriate activity; racism

Ok, I should have been clear in the beginning - what struck me was that the first example was essentially answering the question on doing great harm with minimum spendings - a really wicked "evil EA", I would say. I found it somewhat ironic.

EM, Effective Malevolence

Did you intend to refer to page 83 rather than 82?

I see it's indeed page 83 in the document on arxiv; it was 82 in the pdf on OpenAI website

Is it a pure coincidence that 3 prominent LLMs are announced on the same day?

Naively, maybe they each thought Pi day (March 14th) would get them more attention? I'd guess it's most likely a coincidence given how many big releases there have been recently, but would be amusing if it was Pi day related.

The other alternative was that there was some coordination about releasing LLM. Plenty of people argue that they somehow should coordinate, so it would not be surprising if they actually did it. 

Particular ChatGPT failure mode that I am wondering if GPT-4 passes: routing questions (the ones I tried "can I drive from Boston to Portland, Maine without passing through New Hampshire", "I want to look at the Arctic Ocean from behind my windshield. Can I do this?" ChatGPT was able to answer both <1/10 times). Anyone with access want to try this?

I don't have GPT-4 access, but I was curious how GPT-3.5 got the first question wrong. I just tried it:

Me: can I drive from Boston to Portland, Maine without passing through New Hampshire?

GPT-3.5: Yes, you can drive from Boston to Portland, Maine without passing through New Hampshire by taking the coastal route. To do this, take I-93 North out of Boston and then take Exit 37A for I-95 North. Follow I-95 North until Exit 44 for I-295 North towards South Portland. Continue on I-295 North until Exit 7 for Route 1 North. Follow Route 1 North until you reach Portland. This route will take you along the coast and bypass New Hampshire. However, please note that this route may take longer than driving through New Hampshire on I-93.

It gets the bottom line correct, but the details are completely wrong.

3.5 will be reasonably well distributed between the wrong answer (no) and the right answer + a routing that passes directly through NH. My single Poe  GPT-4 fell into the second category.

There's the claim that GPT-4 is better at not going off the guardrails and that Bing runs on GPT-4. How does that fit together with Bing's behavior?

I think it's referring to the version of GPT-4 with RLHF, which I believe Bing/Sydney doesn't have? Bing/Sydney being based on the pre-trained version or the fine-tuned version, most likely.

Curated and popular this week
Paul Present
 ·  · 28m read
 · 
Note: I am not a malaria expert. This is my best-faith attempt at answering a question that was bothering me, but this field is a large and complex field, and I’ve almost certainly misunderstood something somewhere along the way. Summary While the world made incredible progress in reducing malaria cases from 2000 to 2015, the past 10 years have seen malaria cases stop declining and start rising. I investigated potential reasons behind this increase through reading the existing literature and looking at publicly available data, and I identified three key factors explaining the rise: 1. Population Growth: Africa's population has increased by approximately 75% since 2000. This alone explains most of the increase in absolute case numbers, while cases per capita have remained relatively flat since 2015. 2. Stagnant Funding: After rapid growth starting in 2000, funding for malaria prevention plateaued around 2010. 3. Insecticide Resistance: Mosquitoes have become increasingly resistant to the insecticides used in bednets over the past 20 years. This has made older models of bednets less effective, although they still have some effect. Newer models of bednets developed in response to insecticide resistance are more effective but still not widely deployed.  I very crudely estimate that without any of these factors, there would be 55% fewer malaria cases in the world than what we see today. I think all three of these factors are roughly equally important in explaining the difference.  Alternative explanations like removal of PFAS, climate change, or invasive mosquito species don't appear to be major contributors.  Overall this investigation made me more convinced that bednets are an effective global health intervention.  Introduction In 2015, malaria rates were down, and EAs were celebrating. Giving What We Can posted this incredible gif showing the decrease in malaria cases across Africa since 2000: Giving What We Can said that > The reduction in malaria has be
Neel Nanda
 ·  · 1m read
 · 
TL;DR Having a good research track record is some evidence of good big-picture takes, but it's weak evidence. Strategic thinking is hard, and requires different skills. But people often conflate these skills, leading to excessive deference to researchers in the field, without evidence that that person is good at strategic thinking specifically. I certainly try to have good strategic takes, but it's hard, and you shouldn't assume I succeed! Introduction I often find myself giving talks or Q&As about mechanistic interpretability research. But inevitably, I'll get questions about the big picture: "What's the theory of change for interpretability?", "Is this really going to help with alignment?", "Does any of this matter if we can’t ensure all labs take alignment seriously?". And I think people take my answers to these way too seriously. These are great questions, and I'm happy to try answering them. But I've noticed a bit of a pathology: people seem to assume that because I'm (hopefully!) good at the research, I'm automatically well-qualified to answer these broader strategic questions. I think this is a mistake, a form of undue deference that is both incorrect and unhelpful. I certainly try to have good strategic takes, and I think this makes me better at my job, but this is far from sufficient. Being good at research and being good at high level strategic thinking are just fairly different skillsets! But isn’t someone being good at research strong evidence they’re also good at strategic thinking? I personally think it’s moderate evidence, but far from sufficient. One key factor is that a very hard part of strategic thinking is the lack of feedback. Your reasoning about confusing long-term factors need to extrapolate from past trends and make analogies from things you do understand better, and it can be quite hard to tell if what you're saying is complete bullshit or not. In an empirical science like mechanistic interpretability, however, you can get a lot more fe
Ronen Bar
 ·  · 10m read
 · 
"Part one of our challenge is to solve the technical alignment problem, and that’s what everybody focuses on, but part two is: to whose values do you align the system once you’re capable of doing that, and that may turn out to be an even harder problem", Sam Altman, OpenAI CEO (Link).  In this post, I argue that: 1. "To whose values do you align the system" is a critically neglected space I termed “Moral Alignment.” Only a few organizations work for non-humans in this field, with a total budget of 4-5 million USD (not accounting for academic work). The scale of this space couldn’t be any bigger - the intersection between the most revolutionary technology ever and all sentient beings. While tractability remains uncertain, there is some promising positive evidence (See “The Tractability Open Question” section). 2. Given the first point, our movement must attract more resources, talent, and funding to address it. The goal is to value align AI with caring about all sentient beings: humans, animals, and potential future digital minds. In other words, I argue we should invest much more in promoting a sentient-centric AI. The problem What is Moral Alignment? AI alignment focuses on ensuring AI systems act according to human intentions, emphasizing controllability and corrigibility (adaptability to changing human preferences). However, traditional alignment often ignores the ethical implications for all sentient beings. Moral Alignment, as part of the broader AI alignment and AI safety spaces, is a field focused on the values we aim to instill in AI. I argue that our goal should be to ensure AI is a positive force for all sentient beings. Currently, as far as I know, no overarching organization, terms, or community unifies Moral Alignment (MA) as a field with a clear umbrella identity. While specific groups focus individually on animals, humans, or digital minds, such as AI for Animals, which does excellent community-building work around AI and animal welfare while