{"id":20107,"date":"2023-07-20T04:49:48","date_gmt":"2023-07-20T04:49:48","guid":{"rendered":"https:\/\/nftandcrypto-news.com\/crypto\/chatgpts-capabilities-are-getting-worse-with-age-new-study-claims\/"},"modified":"2023-07-20T04:49:49","modified_gmt":"2023-07-20T04:49:49","slug":"chatgpts-capabilities-are-getting-worse-with-age-new-study-claims","status":"publish","type":"post","link":"https:\/\/nftandcrypto-news.com\/crypto\/chatgpts-capabilities-are-getting-worse-with-age-new-study-claims\/","title":{"rendered":"ChatGPT\u2019s capabilities are getting worse with age, new study claims"},"content":{"rendered":"

<\/p>\n

\n

OpenAI\u2019s artificial intelligence-powered chatbot ChatGPT seems to be getting worse as time goes on and researchers can\u2019t seem to figure out the reason why.\u00a0<\/p>\n

In a July 18 study, researchers from Stanford and UC Berkeley found ChatGPT\u2019s newest models had become far less capable of providing accurate answers to an identical series of questions within the span of a few months. <\/p>\n

The study\u2019s authors couldn\u2019t provide a clear answer as to why the AI chatbot\u2019s capabilities had deteriorated. <\/p>\n

To test how reliable the different models of ChatGPT were, researchers Lingjiao Chen, Matei Zaharia and James Zou asked ChatGPT-3.5 and ChatGPT-4 models to solve a series of math problems, answer sensitive questions, write new lines of code and conduct spatial reasoning from prompts.<\/p>\n

\n

We evaluated #ChatGPT<\/a>‘s behavior over time and found substantial diffs in its responses to the *same questions* between the June version of GPT4 and GPT3.5 and the March versions. The newer versions got worse on some tasks. w\/ Lingjiao Chen @matei_zaharia<\/a> https:\/\/t.co\/TGeN4T18Fd https:\/\/t.co\/36mjnejERy pic.twitter.com\/FEiqrUVbg6<\/a><\/p>\n

\u2014 James Zou (@james_y_zou) July 19, 2023<\/a><\/p><\/blockquote>\n

According to the research, in March ChatGPT-4 was capable of identifying prime numbers with a 97.6% accuracy rate. In the same test conducted in June, GPT-4\u2019s accuracy had plummeted to just 2.4%. <\/p>\n

In contrast, the earlier GPT-3.5 model had improved on prime number identification within the same time frame.<\/p>\n

Related: <\/em><\/strong>SEC\u2019s Gary Gensler believes AI can strengthen its enforcement regime<\/em><\/strong><\/p>\n

When it came to generating lines of new code, the abilities of both models deteriorated substantially between March and June.<\/p>\n

The study also found ChatGPT\u2019s responses to sensitive questions \u2014 with some examples showing a focus on ethnicity and gender \u2014 later became more concise in refusing to answer. <\/p>\n

Earlier iterations of the chatbot provided extensive reasoning for why it couldn\u2019t answer certain sensitive questions. In June however, the models simply apologized to the user and refused to answer.<\/p>\n

\u201cThe behavior of the \u2018same\u2019 [large language model] service can change substantially in a relatively short amount of time,\u201d the researchers wrote, noting the need for continuous monitoring of AI model quality.<\/p>\n

The researchers recommended users and companies who rely on LLM services as a component in their workflows implement some form of monitoring analysis to ensure the chatbot remains up to speed.<\/p>\n

On June 6, OpenAI unveiled plans to create a team that will help manage the risks that could emerge from a superintelligent AI system, something it expects to arrive within the decade.<\/p>\n

AI Eye: <\/em><\/strong>AI\u2019s trained on AI content go MAD, is Threads a loss leader for AI data?<\/em><\/strong><\/p>\n<\/div>\n