Is ChatGPT getting dumber? Evidence suggests yes
A new study from Stanford and UC Berkeley suggests GPT-4's accuracy has declined from 97.6% in March to just 2.4% in June after model updates.
The study demonstrates that GPT-4 now struggles with simple reasoning questions it previously answered correctly, showing no work to explain its (incorrect) answers.
This matches widespread anecdotal observations from users that GPT-4's quality has degraded recently.
The findings contradict OpenAI's claim that complaints are just a psychological effect amongst heavy users. For Open AI’s part, Logan Kilpatrick, who leads developer relations at the organisation, tweeted that they are looking into the report’s findings.
Whilst methodological errors in the study cannot be ruled out, if the findings are true, there will be significant implications for businesses that have relied on GPT-4 to build out additional tools.
Anthropic launches Claude 2 chatbot to rival ChatGPT
American AI company Anthropic has released Claude 2, a new chatbot aimed at competing with ChatGPT, Bard and other AI assistants.
Claude 2 can summarise long passages up to 75,000 words, similar to a novel length. This exceeds ChatGPT's summarisation abilities. It is also able to accept document uploads, something ChatGPT doesn’t do natively.
Claude 2 operates based on a set of principles from documents like the UN Declaration of Human Rights to guide its judgments. Anthropic claims this "Constitutional AI" approach makes it safer.
However, Claude 2 still seems prone to factual errors in its responses, like incorrectly naming sports winners, limiting its capabilities versus rivals.
China introduces new rules for generative AI products
China will introduce interim measures in August to manage its growing generative AI industry, attempting to balance its aim of maintaining security with support for technological development. China is one of the first countries to introduce such regulations.
Firms such as Baidu and Alibaba Group have launched dozens of AI models but held back from rolling out chatbots until new rules were finalised.
The Cyberspace Administration of China said providers who wanted to offer services to the public would need to submit security assessments, ensuring products did not infringe IP rights and used legitimate data sources. Providers must also register their algorithms with the government and ensure content is ‘in line with China's core socialist values’.