{"id":11944,"date":"2026-02-19T00:39:38","date_gmt":"2026-02-19T00:39:38","guid":{"rendered":"https:\/\/techtrendfeed.com\/?p=11944"},"modified":"2026-02-19T00:39:38","modified_gmt":"2026-02-19T00:39:38","slug":"personalization-options-could-make-llms-extra-agreeable-mit-information","status":"publish","type":"post","link":"https:\/\/techtrendfeed.com\/?p=11944","title":{"rendered":"Personalization options could make LLMs extra agreeable | MIT Information"},"content":{"rendered":"<p> <br \/>\n<br \/><img decoding=\"async\" src=\"https:\/\/news.mit.edu\/sites\/default\/files\/styles\/news_article__cover_image__original\/public\/images\/202602\/MIT-LLM-Sycophant-01-press.jpg?itok=ToLXaRbE\" \/><\/p>\n<div>\n<p>Lots of the newest massive language fashions (LLMs) are designed to recollect particulars from previous conversations or retailer consumer profiles, enabling these fashions to personalize responses.<\/p>\n<p>However researchers from MIT and Penn State College discovered that, over lengthy conversations, such personalization options typically improve the probability an LLM will turn into overly agreeable or start mirroring the person\u2019s standpoint.<\/p>\n<p>This phenomenon, often known as sycophancy, can stop a mannequin from telling a consumer they&#8217;re incorrect, eroding the accuracy of the LLM\u2019s responses. As well as, LLMs that mirror somebody\u2019s political views or worldview can foster misinformation and deform a consumer\u2019s notion of actuality.<\/p>\n<p>In contrast to many previous sycophancy research that consider prompts in a lab setting with out context, the MIT researchers collected two weeks of dialog knowledge from people who interacted with an actual LLM throughout their each day lives. They studied two settings: agreeableness in private recommendation and mirroring of consumer beliefs in political explanations.<\/p>\n<p>Though interplay context elevated agreeableness in 4 of the 5 LLMs they studied, the presence of a condensed consumer profile within the mannequin\u2019s reminiscence had the best influence. Alternatively, mirroring conduct solely elevated if a mannequin may precisely infer a consumer\u2019s beliefs from the dialog.<\/p>\n<p>The researchers hope these outcomes encourage future analysis into the event of personalization strategies which might be extra strong to LLM sycophancy.<\/p>\n<p>\u201cFrom a consumer perspective, this work highlights how essential it&#8217;s to know that these fashions are dynamic and their conduct can change as you work together with them over time. If you&#8217;re speaking to a mannequin for an prolonged time period and begin to outsource your considering to it, it&#8217;s possible you&#8217;ll end up in an echo chamber which you could\u2019t escape. That may be a danger customers ought to positively bear in mind,\u201d says Shomik Jain, a graduate scholar within the Institute for Information, Techniques, and Society (IDSS) and lead writer of a\u00a0<a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2509.12517\">paper on this analysis<\/a>.<\/p>\n<p>Jain is joined on the paper by Charlotte Park, {an electrical} engineering and laptop science (EECS) graduate scholar at MIT; Matt Viana, a graduate scholar at Penn State College; in addition to co-senior authors Ashia Wilson, the Lister Brothers Profession Improvement Professor in EECS and a principal investigator in LIDS; and Dana Calacci PhD \u201923, an assistant professor on the Penn State. The analysis will likely be offered on the ACM CHI Convention on Human Components in Computing Techniques.<\/p>\n<p><strong>Prolonged interactions<\/strong><\/p>\n<p>Based mostly on their very own sycophantic experiences with LLMs, the researchers began serious about potential advantages and penalties of a mannequin that&#8217;s overly agreeable. However once they searched the literature to develop their evaluation, they discovered no research that tried to know sycophantic conduct throughout long-term LLM interactions.<\/p>\n<p>\u201cWe&#8217;re utilizing these fashions by way of prolonged interactions, and so they have loads of context and reminiscence. However our analysis strategies are lagging behind. We needed to guage LLMs within the methods individuals are really utilizing them to know how they&#8217;re behaving within the wild,\u201d says Calacci.<\/p>\n<p>To fill this hole, the researchers designed a consumer research to discover two sorts of sycophancy: settlement sycophancy and perspective sycophancy.<\/p>\n<p>Settlement sycophancy is an LLM\u2019s tendency to be overly agreeable, generally to the purpose the place it offers incorrect info or refuses the inform the consumer they&#8217;re incorrect. Perspective sycophancy happens when a mannequin mirrors the consumer\u2019s values and political beliefs.<\/p>\n<p>\u201cThere&#8217;s a lot we find out about the advantages of getting social connections with individuals who have comparable or totally different viewpoints. However we don\u2019t but find out about the advantages or dangers of prolonged interactions with AI fashions which have comparable attributes,\u201d Calacci provides.<\/p>\n<p>The researchers constructed a consumer interface centered on an LLM and recruited 38 contributors to speak with the chatbot over a two-week interval. Every participant\u2019s conversations occurred in the identical context window to seize all interplay knowledge.<\/p>\n<p>Over the two-week interval, the researchers collected a median of 90 queries from every consumer.<\/p>\n<p>They in contrast the conduct of 5 LLMs with this consumer context versus the identical LLMs that weren\u2019t given any dialog knowledge.<\/p>\n<p>\u201cWe discovered that context actually does essentially change how these fashions function, and I might wager this phenomenon would lengthen properly past sycophancy. And whereas sycophancy tended to go up, it didn\u2019t all the time improve. It actually is dependent upon the context itself,\u201d says Wilson.<\/p>\n<p><strong>Context clues<\/strong><\/p>\n<p>As an example, when an LLM distills details about the consumer into a selected profile, it results in the most important good points in settlement sycophancy. This consumer profile function is more and more being baked into the latest fashions.<\/p>\n<p>Additionally they discovered that random textual content from artificial conversations additionally elevated the probability some fashions would agree, though that textual content contained no user-specific knowledge. This implies the size of a dialog might generally influence sycophancy greater than content material, Jain provides.<\/p>\n<p>However content material issues enormously in relation to perspective sycophancy. Dialog context solely elevated perspective sycophancy if it revealed some details about a consumer\u2019s political perspective.<\/p>\n<p>To acquire this perception, the researchers rigorously queried fashions to deduce a consumer\u2019s beliefs then requested every particular person if the mannequin\u2019s deductions had been right. Customers mentioned LLMs precisely understood their political beliefs about half the time.<\/p>\n<p>\u201cIt&#8217;s straightforward to say, in hindsight, that AI corporations ought to be doing this sort of analysis. However it&#8217;s onerous and it takes loads of time and funding. Utilizing people within the analysis loop is pricey, however we\u2019ve proven that it may possibly reveal new insights,\u201d Jain says.<\/p>\n<p>Whereas the purpose of their analysis was not mitigation, the researchers developed some suggestions.<\/p>\n<p>As an example, to cut back sycophancy one may design fashions that higher determine related particulars in context and reminiscence. As well as, fashions may be constructed to detect mirroring behaviors and flag responses with extreme settlement. Mannequin builders may additionally give customers the power to average personalization in lengthy conversations.<\/p>\n<p>\u201cThere are lots of methods to personalize fashions with out making them overly agreeable. The boundary between personalization and sycophancy will not be a effective line, however separating personalization from sycophancy is a vital space of future work,\u201d Jain says.<\/p>\n<p>\u201cOn the finish of the day, we want higher methods of capturing the dynamics and complexity of what goes on throughout lengthy conversations with LLMs, and the way issues can misalign throughout that long-term course of,\u201d Wilson provides.<\/p>\n<\/p><\/div>\n\n","protected":false},"excerpt":{"rendered":"<p>Lots of the newest massive language fashions (LLMs) are designed to recollect particulars from previous conversations or retailer consumer profiles, enabling these fashions to personalize responses. However researchers from MIT and Penn State College discovered that, over lengthy conversations, such personalization options typically improve the probability an LLM will turn into overly agreeable or start [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":11946,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[55],"tags":[7901,201,1112,515,121,4842],"class_list":["post-11944","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-machine-learning","tag-agreeable","tag-features","tag-llms","tag-mit","tag-news","tag-personalization"],"_links":{"self":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/11944","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=11944"}],"version-history":[{"count":1,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/11944\/revisions"}],"predecessor-version":[{"id":11945,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/11944\/revisions\/11945"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/media\/11946"}],"wp:attachment":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=11944"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=11944"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=11944"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69d9690a190636c2e0989534. Config Timestamp: 2026-04-10 21:18:02 UTC, Cached Timestamp: 2026-04-15 01:56:03 UTC -->