Skip to content
Longterm Wiki
Back

Announcing Claude 2

web

Credibility Rating

4/5
High(4)

High quality. Established institution or organization with editorial oversight and accountability.

Rating inherited from publication venue: Anthropic

Anthropic's announcement of Claude 2 details capability improvements alongside safety enhancements, including 2x reduction in harmful outputs and red-teaming results, making it relevant to AI safety deployment practices.

Metadata

Importance: 52/100press releasenews

Summary

Anthropic announces Claude 2, featuring improved performance on coding, math, and reasoning benchmarks, a 100K token context window, and significant safety improvements. The model scored 2x better on harmless responses compared to Claude 1.3 in internal red-team evaluations. It is made available via API and the public claude.ai beta.

Key Points

  • Claude 2 scored 76.5% on the Bar exam multiple choice section (up from 73.0% with Claude 1.3) and above 90th percentile on GRE reading/writing.
  • Supports up to 100K token input context, enabling processing of hundreds of pages of documentation or entire books.
  • Coding performance improved to 71.2% on Codex HumanEval (up from 56.0%) and 88.0% on GSM8k math problems.
  • Internal red-team evaluation showed Claude 2 was 2x better at giving harmless responses compared to Claude 1.3.
  • Safety improvements achieved through red-teaming, Constitutional AI techniques, and iterative safety-focused training.

Cited by 1 page

PageTypeQuality
RLHFResearch Area63.0

Cached Content Preview

HTTP 200Fetched Apr 24, 20266 KB
Announcements Claude 2

 Jul 11, 2023 Talk to Claude We are pleased to announce Claude 2 , our new model. Claude 2 has improved performance, longer responses, and can be accessed via API as well as a new public-facing beta website, claude.ai . We have heard from our users that Claude is easy to converse with, clearly explains its thinking, is less likely to produce harmful outputs, and has a longer memory. We have made improvements from our previous models on coding, math, and reasoning. For example, our latest model scored 76.5% on the multiple choice section of the Bar exam, up from 73.0% with Claude 1.3. When compared to college students applying to graduate school, Claude 2 scores above the 90th percentile on the GRE reading and writing exams, and similarly to the median applicant on quantitative reasoning.

 Think of Claude as a friendly, enthusiastic colleague or personal assistant who can be instructed in natural language to help you with many tasks. The Claude 2 API for businesses is being offered for the same price as Claude 1.3. Additionally, anyone in the US and UK can start using our beta chat experience today.

 As we work to improve both the performance and safety of our models, we have increased the length of Claude’s input and output. Users can input up to  100K tokens in each prompt, which means that Claude can work over hundreds of pages of technical documentation or even a book. Claude can now also write longer documents - from memos to letters to stories up to a few thousand tokens - all in one go.

 

 

 In addition, our latest model has greatly improved coding skills. Claude 2 scored a 71.2% up from 56.0% on the Codex HumanEval , a Python coding test. On GSM8k, a large set of grade-school math problems, Claude 2 scored 88.0% up from 85.2%. We have an exciting roadmap of capability improvements planned for Claude 2 and will be slowly and iteratively deploying them in the coming months.

 

 

 We've been iterating to improve the underlying safety of Claude 2, so that it is more harmless and harder to prompt to produce offensive or dangerous output. We have an internal red-teaming evaluation that scores our models on a large representative set of harmful prompts, using an automated test while we also regularly check the results manually. In this evaluation, Claude 2 was 2x better at giving harmless responses compared to Claude 1.3. Although no model is immune from jailbreaks, we’ve used a variety of safety techniques (which you can read about here and here ), as well as extensive red-teaming , to improve its outputs.

 Claude 2 powers our chat experience, and is generally available in the US and UK. We are working to make Claude more globally available in the coming months. You can now create an account and start talking to Claude in natural language, asking it for help with any tasks that you like. Talking to an AI assistant can take some trial and error, so read up on our tips to get the most out of Claude.

 We are al

... (truncated, 6 KB total)
Resource ID: 39071a9443e5c8dd | Stable ID: sid_aQqkljA7mg