如果你最近让ChatGPT做某件事情却被它拒绝,或者它傲慢地推诿责任,不只你有这样的遭遇。
去年12月,ChatGPT在X上承认,人们担心它变得太懒惰。当时,奥特曼写道“我们从11月11日以后就没有更新模型,这绝非有意为之”,这让数以十亿计依赖这款软件写邮件和编写代码的上班族(和学生)松了一口气。
OpenAI的亿万富翁CEO山姆·奥特曼表示,ChatGPT在实现新年计划的时候起步缓慢,因为它表现出越来越多坏习惯,例如懒惰的倾向,坚持不做它能做的事情,甚至变得有些傲慢。正如奥特曼近日在X上宣布软件更新时承认的那样,ChatGPT机器人“现在应该变得更懒惰了”。
但如果你看过奥特曼最近几个月的评论,你会发现这款机器人的个性(或者个性的幻觉)很难抑制。他在12月发帖称:“模型的行为无法预测。”他补充道,训练聊天机器人模型可能像训练宠物狗一样,它们都不会始终给出相同的反应。
奥特曼写道:“即使使用相同数据集进行的不同训练,可能使模型出现明显不同的个性、书写风格、拒绝行为、评价表现甚至政治偏好。”
OpenAI并未回应置评请求。
ChatGPT的行为变化
在去年年末,越来越多人对这款对话AI平台表示不满,用户抱怨它给出的回答与他们的要求不匹配。在Reddit上,有数十名用户描述了这款聊天机器人在执行任务时的不服从性,例如用户要求它写一篇与一个主题有关的简短总结,却收到了一篇论文。一位Reddit用户写道:“如果我说‘再短一些,用最多三四句话’,它会给我有十句话的一个段落。”
甚至在去年早些时候,就已经有一些警告信号。例如,斯坦福大学/加州大学伯克利分校的一项研究,其名称《ChatGPT的行为正在如何发生变化?》(How is ChatGPT’s behavior changing over time?)可谓恰如其分。研究发现,AI软件处理某些任务的准确度有起伏或巨大波动,在短短三个月后,它解决数学问题和识别素数的准确度,从约97%下降到不足3%。
ChatGPT用户还曝出了其他问题,例如这款机器人会说在2023年最近的更新之后它无法浏览互联网。在用户鼓励它说“你可以的”时,这款聊天机器人似乎记起了它能浏览互联网。Reddit上讨论的话题包括人们是否认为较早版本的软件运行效果更好,还有人认为这款软件的新颖性引起的狂热,让人们误以为它的旧版本比现在的版本更机智。
开发者罗伯·林奇表示,这款聊天机器人在冬季放慢节奏,可能是因为它预期人类也会在节日期间放慢速度,也有可能这是另外一个无法解释的故障。林奇在X上发文称,他对这款聊天机器人进行了测试,当它以为时间是12月而不是5月时,就会给出更简短的回答。
自2022年11月ChatGPT公开发布以来,它已经拥有超过17亿用户。人们借助这款软件写邮件、信函和报告,而更精通技术的用户则用它来寻找程序漏洞,编写代码和进行研究分析。
上班族利用这款软件增强自己的工作能力,并声称它帮助他们“更聪明地工作,而不是工作得更辛苦”。
令人担忧的是,研究发现了聊天机器人在评估某些职业和行为时表现出的性别偏见,它会将男性与医生和上班关联在一起,而与女性关联的则是护士和烹饪。
在客户服务和医疗保健等领域,行业专家警告,过度依赖ChatGPT可能意味着失去一些领域依旧需要培养的人与人之间的联系和同理心。如果你与客户服务机器人通了40分钟电话,你就能深有体会。但现在我们不得不等着看看ChatGPT变懒惰的问题能否解决。(财富中文网)
翻译:刘进龙
审校:汪皓
如果你最近让ChatGPT做某件事情却被它拒绝,或者它傲慢地推诿责任,不只你有这样的遭遇。
去年12月,ChatGPT在X上承认,人们担心它变得太懒惰。当时,奥特曼写道“我们从11月11日以后就没有更新模型,这绝非有意为之”,这让数以十亿计依赖这款软件写邮件和编写代码的上班族(和学生)松了一口气。
OpenAI的亿万富翁CEO山姆·奥特曼表示,ChatGPT在实现新年计划的时候起步缓慢,因为它表现出越来越多坏习惯,例如懒惰的倾向,坚持不做它能做的事情,甚至变得有些傲慢。正如奥特曼近日在X上宣布软件更新时承认的那样,ChatGPT机器人“现在应该变得更懒惰了”。
但如果你看过奥特曼最近几个月的评论,你会发现这款机器人的个性(或者个性的幻觉)很难抑制。他在12月发帖称:“模型的行为无法预测。”他补充道,训练聊天机器人模型可能像训练宠物狗一样,它们都不会始终给出相同的反应。
奥特曼写道:“即使使用相同数据集进行的不同训练,可能使模型出现明显不同的个性、书写风格、拒绝行为、评价表现甚至政治偏好。”
OpenAI并未回应置评请求。
ChatGPT的行为变化
在去年年末,越来越多人对这款对话AI平台表示不满,用户抱怨它给出的回答与他们的要求不匹配。在Reddit上,有数十名用户描述了这款聊天机器人在执行任务时的不服从性,例如用户要求它写一篇与一个主题有关的简短总结,却收到了一篇论文。一位Reddit用户写道:“如果我说‘再短一些,用最多三四句话’,它会给我有十句话的一个段落。”
甚至在去年早些时候,就已经有一些警告信号。例如,斯坦福大学/加州大学伯克利分校的一项研究,其名称《ChatGPT的行为正在如何发生变化?》(How is ChatGPT’s behavior changing over time?)可谓恰如其分。研究发现,AI软件处理某些任务的准确度有起伏或巨大波动,在短短三个月后,它解决数学问题和识别素数的准确度,从约97%下降到不足3%。
ChatGPT用户还曝出了其他问题,例如这款机器人会说在2023年最近的更新之后它无法浏览互联网。在用户鼓励它说“你可以的”时,这款聊天机器人似乎记起了它能浏览互联网。Reddit上讨论的话题包括人们是否认为较早版本的软件运行效果更好,还有人认为这款软件的新颖性引起的狂热,让人们误以为它的旧版本比现在的版本更机智。
开发者罗伯·林奇表示,这款聊天机器人在冬季放慢节奏,可能是因为它预期人类也会在节日期间放慢速度,也有可能这是另外一个无法解释的故障。林奇在X上发文称,他对这款聊天机器人进行了测试,当它以为时间是12月而不是5月时,就会给出更简短的回答。
自2022年11月ChatGPT公开发布以来,它已经拥有超过17亿用户。人们借助这款软件写邮件、信函和报告,而更精通技术的用户则用它来寻找程序漏洞,编写代码和进行研究分析。
上班族利用这款软件增强自己的工作能力,并声称它帮助他们“更聪明地工作,而不是工作得更辛苦”。
令人担忧的是,研究发现了聊天机器人在评估某些职业和行为时表现出的性别偏见,它会将男性与医生和上班关联在一起,而与女性关联的则是护士和烹饪。
在客户服务和医疗保健等领域,行业专家警告,过度依赖ChatGPT可能意味着失去一些领域依旧需要培养的人与人之间的联系和同理心。如果你与客户服务机器人通了40分钟电话,你就能深有体会。但现在我们不得不等着看看ChatGPT变懒惰的问题能否解决。(财富中文网)
翻译:刘进龙
审校:汪皓
If you asked ChatGPT to do something lately and it refused, or sassily gave you the runaround, you’re not alone.
On X last December, ChatGPT acknowledged that people were concerned it’s gotten too lazy. Last December, Altman wrote that “we haven’t updated the model since Nov. 11, and this certainly isn’t intentional,” much to the relief of billions of workers (and students) who have come to rely on the software to write emails and code.
ChatGPT has had a slow start to its New Year’s resolution, according to OpenAI’s billionaire CEO Sam Altman, as it’s displayed an increasing number of bad habits, from a tendency to be lazy to an insistence it can’t do things that it can to even just being kind of sassy. As Altman acknowledged yesterday on X, along with the news of a software update, the ChatGPT bot “should now be much less lazy.”
But if you look at Altman’s comments for months now, the bot’s personalities (or illusions of them) are hard to suppress. “Model behavior can be unpredictable,” he wrote in a December post. Training chat models, he added, might be more similar to training puppies—they don’t always respond the same.
“Different training runs even using the same datasets can produce models that are noticeably different in personality, writing style, refusal behavior, evaluation performance, and even political bias,” Altman wrote.
OpenAI did not respond to a request for comment.
ChatGPT’s changing behavior
As last year wound down, complaints against the conversational AI platform rose as users reported responses that didn’t match their requests. On Reddit, dozens of users described the bot’s disobedience on tasks, like asking it to write a short summary on a topic and instead receiving an essay. One Redditor wrote, “And if i say, ‘much shorter, 3-4 sentences max,’ it’ll give me a paragraph with ten sentences.’”
Warning signs emerged even earlier last year, notably including a Stanford/Berkeley study, aptly named “How is ChatGPT’s behavior changing over time?” It found that the AI software had drifts, or wild fluctuations in accuracy with certain tasks, and its facility in solving math problems and identifying prime numbers had dropped from about 97% to under 3% accuracy after just three months.
ChatGPT users have described other issues, like the bot saying it can’t browse the internet beyond its recent update in 2023. When the user encouraged it, typing in “yes you can,” the bot seemed to remember it could browse the internet. Reddit discussions include whether people think older versions of the software worked better, while others think that the frenzy over the newness of the software makes it seem like it used to be more tactful than now.
One developer, Rob Lynch, posited that the chatbot may have been taking it slow for the winter, possibly due to its expectation that its human counterparts also slow down during the holidays, or just another unexplained glitch. Lynch posted on X that he had run tests on the bot and it would reply in shorter answers when it thought it was December rather than May.
Since ChatGPT’s debut to the public in November 2022, it has amassed an estimate of over 1.7 billion users. People have been using the software to write emails, letters, and reports, and the more tech-savvy depend on it to debug and write codes and research analysis.
Workers have bolstered their manpower with the software and claim it helps them “work smarter, not harder.“
Of concern, studies have revealed the chatbot’s gender bias through its assessment of certain careers and actions, associating men with doctors and going to work, and women to nurses and cooking.
In fields like customer service and health care, industry experts warn that overreliance on ChatGPT could mean a loss of human connection and empathy that some fields still need to foster. If you’ve ever been on a 40-minute call with a customer service robot, you already know. But now we have to wait and see if the puppy’s laziness problem fades.