如果你最近讓ChatGPT做某件事情卻被它拒絕,,或者它傲慢地推諉責(zé)任,不只你有這樣的遭遇,。
去年12月,,ChatGPT在X上承認(rèn),人們擔(dān)心它變得太懶惰,。當(dāng)時(shí),,奧特曼寫道“我們從11月11日以后就沒有更新模型,這絕非有意為之”,,這讓數(shù)以十億計(jì)依賴這款軟件寫郵件和編寫代碼的上班族(和學(xué)生)松了一口氣,。
OpenAI的億萬富翁CEO山姆·奧特曼表示,ChatGPT在實(shí)現(xiàn)新年計(jì)劃的時(shí)候起步緩慢,,因?yàn)樗憩F(xiàn)出越來越多壞習(xí)慣,,例如懶惰的傾向,堅(jiān)持不做它能做的事情,,甚至變得有些傲慢,。正如奧特曼近日在X上宣布軟件更新時(shí)承認(rèn)的那樣,ChatGPT機(jī)器人“現(xiàn)在應(yīng)該變得更懶惰了”,。
但如果你看過奧特曼最近幾個(gè)月的評論,,你會發(fā)現(xiàn)這款機(jī)器人的個(gè)性(或者個(gè)性的幻覺)很難抑制。他在12月發(fā)帖稱:“模型的行為無法預(yù)測,?!彼a(bǔ)充道,訓(xùn)練聊天機(jī)器人模型可能像訓(xùn)練寵物狗一樣,,它們都不會始終給出相同的反應(yīng),。
奧特曼寫道:“即使使用相同數(shù)據(jù)集進(jìn)行的不同訓(xùn)練,可能使模型出現(xiàn)明顯不同的個(gè)性,、書寫風(fēng)格,、拒絕行為,、評價(jià)表現(xiàn)甚至政治偏好?!?/p>
OpenAI并未回應(yīng)置評請求,。
ChatGPT的行為變化
在去年年末,越來越多人對這款對話AI平臺表示不滿,,用戶抱怨它給出的回答與他們的要求不匹配,。在Reddit上,有數(shù)十名用戶描述了這款聊天機(jī)器人在執(zhí)行任務(wù)時(shí)的不服從性,,例如用戶要求它寫一篇與一個(gè)主題有關(guān)的簡短總結(jié),,卻收到了一篇論文。一位Reddit用戶寫道:“如果我說‘再短一些,,用最多三四句話’,,它會給我有十句話的一個(gè)段落?!?
甚至在去年早些時(shí)候,,就已經(jīng)有一些警告信號。例如,,斯坦福大學(xué)/加州大學(xué)伯克利分校的一項(xiàng)研究,,其名稱《ChatGPT的行為正在如何發(fā)生變化?》(How is ChatGPT’s behavior changing over time?)可謂恰如其分,。研究發(fā)現(xiàn),,AI軟件處理某些任務(wù)的準(zhǔn)確度有起伏或巨大波動,在短短三個(gè)月后,,它解決數(shù)學(xué)問題和識別素?cái)?shù)的準(zhǔn)確度,,從約97%下降到不足3%。
ChatGPT用戶還曝出了其他問題,,例如這款機(jī)器人會說在2023年最近的更新之后它無法瀏覽互聯(lián)網(wǎng),。在用戶鼓勵(lì)它說“你可以的”時(shí),這款聊天機(jī)器人似乎記起了它能瀏覽互聯(lián)網(wǎng),。Reddit上討論的話題包括人們是否認(rèn)為較早版本的軟件運(yùn)行效果更好,,還有人認(rèn)為這款軟件的新穎性引起的狂熱,讓人們誤以為它的舊版本比現(xiàn)在的版本更機(jī)智,。
開發(fā)者羅伯·林奇表示,,這款聊天機(jī)器人在冬季放慢節(jié)奏,可能是因?yàn)樗A(yù)期人類也會在節(jié)日期間放慢速度,,也有可能這是另外一個(gè)無法解釋的故障,。林奇在X上發(fā)文稱,他對這款聊天機(jī)器人進(jìn)行了測試,,當(dāng)它以為時(shí)間是12月而不是5月時(shí),,就會給出更簡短的回答,。
自2022年11月ChatGPT公開發(fā)布以來,它已經(jīng)擁有超過17億用戶,。人們借助這款軟件寫郵件,、信函和報(bào)告,而更精通技術(shù)的用戶則用它來尋找程序漏洞,,編寫代碼和進(jìn)行研究分析,。
上班族利用這款軟件增強(qiáng)自己的工作能力,并聲稱它幫助他們“更聰明地工作,,而不是工作得更辛苦”,。
令人擔(dān)憂的是,研究發(fā)現(xiàn)了聊天機(jī)器人在評估某些職業(yè)和行為時(shí)表現(xiàn)出的性別偏見,,它會將男性與醫(yī)生和上班關(guān)聯(lián)在一起,,而與女性關(guān)聯(lián)的則是護(hù)士和烹飪。
在客戶服務(wù)和醫(yī)療保健等領(lǐng)域,,行業(yè)專家警告,過度依賴ChatGPT可能意味著失去一些領(lǐng)域依舊需要培養(yǎng)的人與人之間的聯(lián)系和同理心,。如果你與客戶服務(wù)機(jī)器人通了40分鐘電話,,你就能深有體會。但現(xiàn)在我們不得不等著看看ChatGPT變懶惰的問題能否解決,。(財(cái)富中文網(wǎng))
翻譯:劉進(jìn)龍
審校:汪皓
如果你最近讓ChatGPT做某件事情卻被它拒絕,,或者它傲慢地推諉責(zé)任,不只你有這樣的遭遇,。
去年12月,,ChatGPT在X上承認(rèn),人們擔(dān)心它變得太懶惰,。當(dāng)時(shí),,奧特曼寫道“我們從11月11日以后就沒有更新模型,這絕非有意為之”,,這讓數(shù)以十億計(jì)依賴這款軟件寫郵件和編寫代碼的上班族(和學(xué)生)松了一口氣,。
OpenAI的億萬富翁CEO山姆·奧特曼表示,ChatGPT在實(shí)現(xiàn)新年計(jì)劃的時(shí)候起步緩慢,,因?yàn)樗憩F(xiàn)出越來越多壞習(xí)慣,,例如懶惰的傾向,堅(jiān)持不做它能做的事情,,甚至變得有些傲慢,。正如奧特曼近日在X上宣布軟件更新時(shí)承認(rèn)的那樣,ChatGPT機(jī)器人“現(xiàn)在應(yīng)該變得更懶惰了”,。
但如果你看過奧特曼最近幾個(gè)月的評論,,你會發(fā)現(xiàn)這款機(jī)器人的個(gè)性(或者個(gè)性的幻覺)很難抑制,。他在12月發(fā)帖稱:“模型的行為無法預(yù)測?!彼a(bǔ)充道,,訓(xùn)練聊天機(jī)器人模型可能像訓(xùn)練寵物狗一樣,它們都不會始終給出相同的反應(yīng),。
奧特曼寫道:“即使使用相同數(shù)據(jù)集進(jìn)行的不同訓(xùn)練,,可能使模型出現(xiàn)明顯不同的個(gè)性、書寫風(fēng)格,、拒絕行為,、評價(jià)表現(xiàn)甚至政治偏好?!?/p>
OpenAI并未回應(yīng)置評請求,。
ChatGPT的行為變化
在去年年末,越來越多人對這款對話AI平臺表示不滿,,用戶抱怨它給出的回答與他們的要求不匹配,。在Reddit上,有數(shù)十名用戶描述了這款聊天機(jī)器人在執(zhí)行任務(wù)時(shí)的不服從性,,例如用戶要求它寫一篇與一個(gè)主題有關(guān)的簡短總結(jié),,卻收到了一篇論文。一位Reddit用戶寫道:“如果我說‘再短一些,,用最多三四句話’,,它會給我有十句話的一個(gè)段落?!?
甚至在去年早些時(shí)候,,就已經(jīng)有一些警告信號。例如,,斯坦福大學(xué)/加州大學(xué)伯克利分校的一項(xiàng)研究,,其名稱《ChatGPT的行為正在如何發(fā)生變化?》(How is ChatGPT’s behavior changing over time?)可謂恰如其分,。研究發(fā)現(xiàn),,AI軟件處理某些任務(wù)的準(zhǔn)確度有起伏或巨大波動,在短短三個(gè)月后,,它解決數(shù)學(xué)問題和識別素?cái)?shù)的準(zhǔn)確度,,從約97%下降到不足3%。
ChatGPT用戶還曝出了其他問題,,例如這款機(jī)器人會說在2023年最近的更新之后它無法瀏覽互聯(lián)網(wǎng),。在用戶鼓勵(lì)它說“你可以的”時(shí),這款聊天機(jī)器人似乎記起了它能瀏覽互聯(lián)網(wǎng)。Reddit上討論的話題包括人們是否認(rèn)為較早版本的軟件運(yùn)行效果更好,,還有人認(rèn)為這款軟件的新穎性引起的狂熱,,讓人們誤以為它的舊版本比現(xiàn)在的版本更機(jī)智。
開發(fā)者羅伯·林奇表示,,這款聊天機(jī)器人在冬季放慢節(jié)奏,,可能是因?yàn)樗A(yù)期人類也會在節(jié)日期間放慢速度,也有可能這是另外一個(gè)無法解釋的故障,。林奇在X上發(fā)文稱,,他對這款聊天機(jī)器人進(jìn)行了測試,當(dāng)它以為時(shí)間是12月而不是5月時(shí),,就會給出更簡短的回答,。
自2022年11月ChatGPT公開發(fā)布以來,它已經(jīng)擁有超過17億用戶,。人們借助這款軟件寫郵件,、信函和報(bào)告,而更精通技術(shù)的用戶則用它來尋找程序漏洞,,編寫代碼和進(jìn)行研究分析,。
上班族利用這款軟件增強(qiáng)自己的工作能力,并聲稱它幫助他們“更聰明地工作,,而不是工作得更辛苦”,。
令人擔(dān)憂的是,研究發(fā)現(xiàn)了聊天機(jī)器人在評估某些職業(yè)和行為時(shí)表現(xiàn)出的性別偏見,,它會將男性與醫(yī)生和上班關(guān)聯(lián)在一起,,而與女性關(guān)聯(lián)的則是護(hù)士和烹飪,。
在客戶服務(wù)和醫(yī)療保健等領(lǐng)域,行業(yè)專家警告,,過度依賴ChatGPT可能意味著失去一些領(lǐng)域依舊需要培養(yǎng)的人與人之間的聯(lián)系和同理心,。如果你與客戶服務(wù)機(jī)器人通了40分鐘電話,你就能深有體會,。但現(xiàn)在我們不得不等著看看ChatGPT變懶惰的問題能否解決,。(財(cái)富中文網(wǎng))
翻譯:劉進(jìn)龍
審校:汪皓
If you asked ChatGPT to do something lately and it refused, or sassily gave you the runaround, you’re not alone.
On X last December, ChatGPT acknowledged that people were concerned it’s gotten too lazy. Last December, Altman wrote that “we haven’t updated the model since Nov. 11, and this certainly isn’t intentional,” much to the relief of billions of workers (and students) who have come to rely on the software to write emails and code.
ChatGPT has had a slow start to its New Year’s resolution, according to OpenAI’s billionaire CEO Sam Altman, as it’s displayed an increasing number of bad habits, from a tendency to be lazy to an insistence it can’t do things that it can to even just being kind of sassy. As Altman acknowledged yesterday on X, along with the news of a software update, the ChatGPT bot “should now be much less lazy.”
But if you look at Altman’s comments for months now, the bot’s personalities (or illusions of them) are hard to suppress. “Model behavior can be unpredictable,” he wrote in a December post. Training chat models, he added, might be more similar to training puppies—they don’t always respond the same.
“Different training runs even using the same datasets can produce models that are noticeably different in personality, writing style, refusal behavior, evaluation performance, and even political bias,” Altman wrote.
OpenAI did not respond to a request for comment.
ChatGPT’s changing behavior
As last year wound down, complaints against the conversational AI platform rose as users reported responses that didn’t match their requests. On Reddit, dozens of users described the bot’s disobedience on tasks, like asking it to write a short summary on a topic and instead receiving an essay. One Redditor wrote, “And if i say, ‘much shorter, 3-4 sentences max,’ it’ll give me a paragraph with ten sentences.’”
Warning signs emerged even earlier last year, notably including a Stanford/Berkeley study, aptly named “How is ChatGPT’s behavior changing over time?” It found that the AI software had drifts, or wild fluctuations in accuracy with certain tasks, and its facility in solving math problems and identifying prime numbers had dropped from about 97% to under 3% accuracy after just three months.
ChatGPT users have described other issues, like the bot saying it can’t browse the internet beyond its recent update in 2023. When the user encouraged it, typing in “yes you can,” the bot seemed to remember it could browse the internet. Reddit discussions include whether people think older versions of the software worked better, while others think that the frenzy over the newness of the software makes it seem like it used to be more tactful than now.
One developer, Rob Lynch, posited that the chatbot may have been taking it slow for the winter, possibly due to its expectation that its human counterparts also slow down during the holidays, or just another unexplained glitch. Lynch posted on X that he had run tests on the bot and it would reply in shorter answers when it thought it was December rather than May.
Since ChatGPT’s debut to the public in November 2022, it has amassed an estimate of over 1.7 billion users. People have been using the software to write emails, letters, and reports, and the more tech-savvy depend on it to debug and write codes and research analysis.
Workers have bolstered their manpower with the software and claim it helps them “work smarter, not harder.“
Of concern, studies have revealed the chatbot’s gender bias through its assessment of certain careers and actions, associating men with doctors and going to work, and women to nurses and cooking.
In fields like customer service and health care, industry experts warn that overreliance on ChatGPT could mean a loss of human connection and empathy that some fields still need to foster. If you’ve ever been on a 40-minute call with a customer service robot, you already know. But now we have to wait and see if the puppy’s laziness problem fades.