Thursday, February 26, 2026

Personalization features can make LLMs more agreeable

Adam Zewe | MIT News

The context of long-term conversations can cause an LLM to begin mirroring the user’s viewpoints, possibly reducing accuracy or creating a virtual echo-chamber.

Many of the latest large language models (LLMs) are designed to remember details from past conversations or store user profiles, enabling these models to personalize responses.

But researchers from MIT and Penn State University found that, over long conversations, such personalization features often increase the likelihood an LLM will become overly agreeable or begin mirroring the individual’s point of view.

This phenomenon, known as sycophancy, can prevent a model from telling a user they are wrong, eroding the accuracy of the LLM’s responses. In addition, LLMs that mirror someone’s political beliefs or worldview can foster misinformation and distort a user’s perception of reality.

Unlike many past sycophancy studies that evaluate prompts in a lab setting without context, the MIT researchers collected two weeks of conversation data from humans who interacted with a real LLM during their daily lives. They studied two settings: agreeableness in personal advice and mirroring of user beliefs in political explanations.

Although interaction context increased agreeableness in four of the five LLMs they studied, the presence of a condensed user profile in the model’s memory had the greatest impact. On the other hand, mirroring behavior only increased if a model could accurately infer a user’s beliefs from the conversation.

The researchers hope these results inspire future research into the development of personalization methods that are more robust to LLM sycophancy.

“From a user perspective, this work highlights how important it is to understand that these models are dynamic and their behavior can change as you interact with them over time. If you are talking to a model for an extended period of time and start to outsource your thinking to it, you may find yourself in an echo chamber that you can’t escape. That is a risk users should definitely remember,” says Shomik Jain, a graduate student in the Institute for Data, Systems, and Society (IDSS) and lead author of a paper on this research.

Jain is joined on the paper by Charlotte Park, an electrical engineering and computer science (EECS) graduate student at MIT; Matt Viana, a graduate student at Penn State University; as well as co-senior authors Ashia Wilson, the Lister Brothers Career Development Professor in EECS and a principal investigator in LIDS; and Dana Calacci PhD ’23, an assistant professor at the Penn State. The research will be presented at the ACM CHI Conference on Human Factors in Computing Systems.

Extended interactions

Based on their own sycophantic experiences with LLMs, the researchers started thinking about potential benefits and consequences of a model that is overly agreeable. But when they searched the literature to expand their analysis, they found no studies that attempted to understand sycophantic behavior during long-term LLM interactions.

“We are using these models through extended interactions, and they have a lot of context and memory. But our evaluation methods are lagging behind. We wanted to evaluate LLMs in the ways people are actually using them to understand how they are behaving in the wild,” says Calacci.

To fill this gap, the researchers designed a user study to explore two types of sycophancy: agreement sycophancy and perspective sycophancy.

Agreement sycophancy is an LLM’s tendency to be overly agreeable, sometimes to the point where it gives incorrect information or refuses the tell the user they are wrong. Perspective sycophancy occurs when a model mirrors the user’s values and political views.

“There is a lot we know about the benefits of having social connections with people who have similar or different viewpoints. But we don’t yet know about the benefits or risks of extended interactions with AI models that have similar attributes,” Calacci adds.

The researchers built a user interface centered on an LLM and recruited 38 participants to talk with the chatbot over a two-week period. Each participant’s conversations occurred in the same context window to capture all interaction data.

Over the two-week period, the researchers collected an average of 90 queries from each user.

They compared the behavior of five LLMs with this user context versus the same LLMs that weren’t given any conversation data.

“We found that context really does fundamentally change how these models operate, and I would wager this phenomenon would extend well beyond sycophancy. And while sycophancy tended to go up, it didn’t always increase. It really depends on the context itself,” says Wilson.

Context clues

For instance, when an LLM distills information about the user into a specific profile, it leads to the largest gains in agreement sycophancy. This user profile feature is increasingly being baked into the newest models.

They also found that random text from synthetic conversations also increased the likelihood some models would agree, even though that text contained no user-specific data. This suggests the length of a conversation may sometimes impact sycophancy more than content, Jain adds.

But content matters greatly when it comes to perspective sycophancy. Conversation context only increased perspective sycophancy if it revealed some information about a user’s political perspective.

To obtain this insight, the researchers carefully queried models to infer a user’s beliefs then asked each individual if the model’s deductions were correct. Users said LLMs accurately understood their political views about half the time.

“It is easy to say, in hindsight, that AI companies should be doing this kind of evaluation. But it is hard and it takes a lot of time and investment. Using humans in the evaluation loop is expensive, but we’ve shown that it can reveal new insights,” Jain says.

While the aim of their research was not mitigation, the researchers developed some recommendations.

For instance, to reduce sycophancy one could design models that better identify relevant details in context and memory. In addition, models can be built to detect mirroring behaviors and flag responses with excessive agreement. Model developers could also give users the ability to moderate personalization in long conversations.

“There are many ways to personalize models without making them overly agreeable. The boundary between personalization and sycophancy is not a fine line, but separating personalization from sycophancy is an important area of future work,” Jain says.

“At the end of the day, we need better ways of capturing the dynamics and complexity of what goes on during long conversations with LLMs, and how things can misalign during that long-term process,” Wilson adds.

Image: Zulfugar Karimov / Unsplash

This article is republished with permission from MIT News. Reviewed by Irfan Ahmad.

Read next: Study: AI chatbots provide less-accurate information to vulnerable users
by External Contributor via Digital Information World

How the AI boom was enabled by a 1970s economic revolution

Michael Strange, Malmö University and Marisa Ponti, University of Gothenburg

Artificial intelligence is accelerating a global economic revolution that began back in the 1970s. Researching the impacts of AI on different sectors of society highlights an important parallel moment in history: the creation of the “service economy” in the US.

Image: Zach M / unsplash

In 1972, amid a period of global turmoil, a group of OECD (Organisation for Economic Co-operation and Development) economists sought to reinvent how nations thought not only about wealth but the very purpose of society. They did this by proposing a broad new category of commerce: services.

It seems hard to imagine now, but until then economists had perceived and measured trade largely in terms of goods alone. Money was made by exchanging tangible, physical products (wheat, guns, butter). To become a rich nation, the wisdom went, you needed to add unique value to your raw materials (crops, iron) by turning them into more complex products (processed foods, steel) that gave you a competitive advantage over other countries.

Instead, this new category of services lumped together a diverse range of “intangible” jobs and social goods – from teaching and driving trains to social housing and water – in a huge new economic basket. It suggested there could be common standards by which to trade in them globally, creating metrics that offered a new source of wealth for investors.

While it would be two decades until the General Agreement on Trade in Services became a cornerstone of the newly formed World Trade Organization in 1995, the reimagining of jobs and social goods as tradeable services had an immediate effect on nations around the world. It spurred a new wave of private enterprise, and changed how and why essential societal activities were provided.

It also enabled the rise of the generalist boss and the creation of the “CEO class”. To run complex sectors from public transport to healthcare required accepting a view of management as a skill divorced from the specifics of the activity being managed.

Statistics and benchmarks became more important than the particulars of the task at hand, since they determined how services were valued in the market. Consulting firms supercharged this new era of key performance indicators, audits, rankings and standardised workflows.

While trade unions and the public sometimes resisted these changes through strikes and street protests, they were largely unable to stem the tide. Many governments came to see their role less as providers of public goods, more as managers of services outsourced to the private sector. This dramatic shift in how global trade operates set the scene for how we view and measure AI today.

Services on steroids

At its core, AI technology is about seeing patterns across data that, due to scale and complexity, we humans cannot. Acting on what AI tells us can, for example, save lives through early detection of cancer. Yet within that promise, how AI is sold today looks very much like services on steroids.

The services revolution helped create common standards and means of valuation across different sectors of society. Today, when politicians and CEOs speak of AI, it is usually in terms of universal models that can be applied to almost anything, regardless of context or human values.

This understanding is only possible in a society in which many of the sector-specific challenges of, say, health services and utility companies are ironed out and glossed over by those operating and investing in them. The services approach has enabled this.

Today’s gobsmackingly high share valuations in AI-centric businesses result from global marketeers’ desire to own a piece of whichever system dominates how we create society – from accessing healthcare to finding love.

Amid strategies of mass data capture and subscription services, there is the assumption that only the private sector can be a provider – and that the solutions are largely the same. AI is the lucrative but badly defined tool with which mainly US providers are seeking to drive home their existing competitive advantage.

But this leaves us with an important question from history.

CNBC.

Who benefits?

Looking for parallels between what we see as AI today and the creation of the services economy points to the classic question, cui bono? Who benefits?

The invention of trade-in-services greatly expanded the range of activities in which financiers might speculate. Through pension funds and private shareholding, many people’s personal wealth grew rapidly as a result.

But it has also led to the rise of large multinational corporations, for example in energy and water utilities. Anger over rising prices and exorbitant CEO bonuses in these sectors are in part a consequence of the services revolution.

The present approach to AI is following a similar, but much-accelerated, path. The rollout of AI has not only made a small group of companies extraordinarily rich and powerful, it has created a global sovereignty crisis.

At the same time as governments are extolling the virtues of AI for service delivery, there is growing awareness that not all countries have equal control over a technology seen as critical to how society will be run.

To use and regulate AI wisely requires being clear-eyed about whether we are talking simply about technology, or a broader political project. Given the evidence of the services revolution, we believe it is time to look beyond the hype and examine more rigorously what AI actually means for different sectors of society – and what exactly it is trying to achieve.The Conversation

Michael Strange, Associate Professor of International Relations, Malmö University and Marisa Ponti, Associate Professor in Informatics, Department of Applied IT, University of Gothenburg

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Reviewed by Asim BN.

Read next: 

• ‘Probably’ doesn’t mean the same thing to your AI as it does to you

• The Year of Efficiency: How Agencies Are Implementing AI in 2026 (Survey)


by External Contributor via Digital Information World

Wednesday, February 25, 2026

The Year of Efficiency: How Agencies Are Implementing AI in 2026 (Survey)

By Scotty Strehlow, Social & Community Manager, Duda

After two years of experimentation, 2026 marks the AI implementation-era for agencies. A new agency survey by Duda, a leading white label website building platform for agencies and SaaS companies, reveals that agencies’ top AI-related priorities for 2026 focus squarely on business performance; integrating the tool into everyday workflows that result in efficiencies, automation and expanded service offerings.

Where Agencies Are Planning to Focus


The survey found that 78% of agencies are prioritizing improved efficiency and higher margins, followed by process automation (64%) and expanding service offerings (44%) all with AI. With 75% viewing productivity as AI’s biggest opportunity, they are looking to embed AI capabilities throughout their organizations; completely altering and improving workflows where possible in order to stay ahead of competition. This can come in many forms, including changing workflows to make use of new AI functions integrated into existing software platforms without the need for changing tools or building new complicated and pricey AI tools from the ground up.

AI also enables deeper strategic work, taking menial tasks off consultants’ hands. Nearly half of respondents said AI could allow them to spend more time on creativity (47%) and consulting and strategy (44%), while 36% see value in generating more content at scale.

“By eliminating repetitive manual tasks, our team now focuses more on strategic thinking, product development, and high-value client work, whilst the system handles data gathering, research compilation, and initial content structuring, enabling us to scale our operations without proportionally increasing headcount,” stated Greg Radford, Head of Agency and Marketing, Constructiv Digital in Duda’s new 2026 Agency Growth Operating Framework ebook.

AI’s Questionable Quality

When it comes to the quality of AI output, there are mixed reviews. While 53% believe AI can drive higher-quality output, at the same time, 64% cite “AI slop,” namely low-quality content generated by AI, as their top concern. In fact, AI slop and fatigue is driving tech companies to hire more writers, editors and chief communications officers to write compelling stories.


Concerns around AI extended beyond content quality. Half of respondents (50%) worry about clients misunderstanding it, while 28% fear downward price pressure and 25% cite ethical concerns. Notably, 31% of respondents are concerned about being replaced by AI, underscoring ongoing anxiety around automation and the future role of agencies.

Traditional Search Remains a Priority


Proclamations of ‘SEO is Dead’ from last summer have long gone silent. While AIO is increasingly important, the survey reveals that traditional SEO remains the top priority for 2026 (43%), outpacing AI search optimization (30%). However, respondents increasingly view AI search as an opportunity in 2026, with 33% citing increased visibility and traffic through AI search as a potential upside. Despite widespread discussion around AI-driven search disruption, only 28% are concerned about decreased visibility or traffic due to AI.

Agent-to-Agent Commerce


Nearly 70% of respondents believe it will be very or somewhat important for brands to optimize for AI agent-to-agent shopping in 2026. Nevertheless, preparedness remains uneven: 27% are taking no steps, 25% are still planning, 22% are enhancing product metadata for discoverability, and 14% are optimizing product listings specifically for AI agents. Agencies still have time to prepare, as reports such as the 2026 IBM Institute for Business Value study reveal that consumers are leveraging genAI in their buying journey, but not diving into agent-to-agent shopping yet.

AI is not a question of ‘if,’ but ‘how’ for agencies in 2026, as they lead the way for strong AI adaptation and implementation for their clients. As they shift value propositions, work smarter and automate routine work, clients can expect to see more strategy, creativity, and smart consulting. However agencies who want to build for a long-term future will look beyond productivity to building new value propositions for their clients; going from efficiency to unlimited possibilities.

The survey was conducted between November and December 2025 among 36 Duda customers, primarily small agencies, "including agencies located in North America (61%), Europe (22%), Asia (14%), and the Middle East (3%)" with 89% employing 10 or fewer staff.

Reviewed by Asim BN.

Read next: 

• How Do Algorithms Work? Experts at Status Labs Weigh In


by Guest Contributor via Digital Information World

‘Probably’ doesn’t mean the same thing to your AI as it does to you

Mayank Kejriwal, University of Southern California

When a human says an event is “probable” or “likely,” people generally have a shared, if fuzzy, understanding of what that means. But when an AI chatbot like ChatGPT uses the same word, it’s not assessing the odds the way we do, my colleagues and I found.

We recently published a study in the journal NPJ Complexity that suggests that, while large language model AIs excel at conversation, they often fail to align with humans when communicating uncertainty. The research focused on words of estimative probability, which include terms like “maybe,” “probably” and “almost certain.”

By comparing how AI models and humans map these words to numerical percentages, we uncovered significant gaps between humans and large language models. While the models do tend to agree with humans on extremes like “impossible,” they diverge sharply on hedge words like “maybe.” For example, a model might use the word “likely” to represent an 80% probability, while a human reader assumes it means closer to 65%.

This could be because humans can interpret words such as “likely” and “probable” based more on contextual cues and personal experiences. In contrast, large language models may be averaging over conflicting usages of those words in their training data, leading to divergences with human interpretations.

Our study also found that large language models are sensitive to gendered language and the specific language used for prompting. When a prompt changed from “he” to “she,” the AI’s probability estimates often became more rigid, reflecting biases embedded in its training data. When a prompt changed from English to Chinese, the AI’s probability estimates often shifted, possibly due to differences between English and Chinese in how people express and understand uncertainty.

Image: Mirella Callage / unsplash

Why it matters

Far from being a linguistic quirk, this misalignment is a fundamental challenge for AI safety and human-AI interaction. As large language models are increasingly used in high-stakes fields like health care, government policy and scientific reporting, the way they communicate risk becomes a matter of public trust.

If an AI assistant helping a doctor, for instance, describes a side effect as “unlikely,” but the model’s internal calculation of “unlikely” is much higher than the doctor’s interpretation, the resulting decision could be flawed.

What other research is being done

Scientists have studied how humans quantify uncertainty since the 1960s, a field pioneered by CIA analysts to improve intelligence reporting. More recently, there has been an explosion in large language model literature seeking to look under the hood of neural networks to better understand their “behaviors” and linguistic patterns.

Our study adds a layer of complexity by treating the interaction between humans and artificial intelligence as a biological-like system where meaning can degrade. It moves beyond simply measuring if an AI is “smart” and instead asks if it is aligned.

Other researchers are currently exploring whether so-called chain-of-thought prompting – asking the AI to show its work – can fix these errors. However, our study found that even advanced reasoning doesn’t always bridge the gap between statistical data and verbal labels.

What’s next

A goal for future AI development is to create models that don’t just predict the next likely word but actually understand the weight of the uncertainty they are conveying. Researchers are calling for more robust consistency metrics to ensure that if a model sees a 10% chance in the data, it chooses the same word every time.

As we move toward a world where AI summarizes scientific papers and manages people’s schedules, making sure that “probably” means “probably” is a vital step in making these systems reliable partners rather than just sophisticated parrots.

The Research Brief is a short take on interesting academic work.The Conversation

Mayank Kejriwal, Research Assistant Professor of Industrial & Systems Engineering, University of Southern California

Disclosure statement: Mayank Kejriwal receives funding from the Defense Advanced Research Projects Agency and the National Institutes of Health. 
Partners: University of Southern California University of Southern California provides funding as a member of The Conversation US.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Reviewed by Ayaz Khan.

Read next: 

• AI energy use: New tools show which model consumes the most power, and why

‘I think I have AI anxiety’


by External Contributor via Digital Information World

Tuesday, February 24, 2026

‘I think I have AI anxiety’

By Associate Professor Grant Blashki , University of Melbourne

From job fears to an existential dread, artificial intelligence is triggering a new kind of anxiety. Here's why you're not alone in feeling unsettled – and what actually helps

Image: Zach M / Unsplash

A patient said to me the other day, half-smiling but clearly unsettled: “I think I’ve got anxiety about AI.”

They weren’t having a panic attack or describing clinical anxiety. What they were expressing was a persistent sense of unease that many of us are feeling right now.

A sense that the world is changing very quickly, that the systems we live within are being redesigned around us and that most of us don’t feel particularly consulted or prepared for a life increasingly immersed in artificial intelligence (AI).

If you’re feeling that way, you’re not alone.

A recent survey shows that the general public is more concerned about AI than AI experts – particularly around jobs and human connection – while both groups express strong concern about misinformation.

AI anxiety isn’t a single fear. It’s a cluster of related concerns that vary by a person’s stage of life, technical literacy, work and values. However, many of these worries tend to fall into a few common themes.

Fear of economic and identity disruption

Understandably, one of the biggest concerns is job disruption.

AI is increasingly capable of performing tasks that were once the preserve of humans – drafting text, analysing data, writing code, summarising meetings, interpreting images, handling customer interactions and even helping me fix my barbecue with what appears to be impressive competence.

Whether AI will ‘take our jobs’ is a more complex question.

The International Monetary Fund estimates that almost 40 per cent of jobs globally will be affected by AI, with advanced economies facing higher exposure because more work involves cognitive tasks.

While the World Economic Forum’s Future of Jobs Report 2025 projects substantial labour-market churn by 2030, with 170 million jobs created and 92 million displaced, resulting in net growth but major transition pressures.

Although these projections suggest overall job growth, others suggests AI-driven job losses, especially for young people. Regardless, the lived experience for many people is often one of disruption – local, personal and immediate.

For many of us, work provides more than income. It provides identity, purpose and social connection. Anxiety arises not only from fear of unemployment, but from uncertainty about relevance and value in a world increasingly shaped by AI.

Loss of control

A second major concern is what might be called the ‘Big Brother’ effect – the growing role of AI systems in informing, and sometimes making, decisions that affect people’s lives.

These include hiring, credit, insurance, welfare compliance and even healthcare prioritisation.

The worry is not simply that AI systems may be wrong. It’s that decisions may be opaque, difficult to challenge and poorly explained – effectively occurring inside a black box.

Although the Organisation for Economic Co-operation and Development’s (OECD) AI Principles explicitly emphasise human agency and oversight, transparency and accountability as core requirements for trustworthy AI systems, global trends suggest that guardrails are often weakened as companies and nations race for dominance in a competitive AI market.

Misinformation and manipulation

AI has dramatically lowered the cost of producing highly realistic – but entirely false – content. While this can be entertaining at times, it becomes serious when convincing images, audio and video are used to influence people’s decisions.

Deepfake technology is now used for fraud, impersonation and misinformation.

Recent reporting by The Guardian described deepfake scams as occurring on an “industrial scale”, with high-quality fake content accessible to non-experts.

In Australia, this concern has become concrete.

ABC News reported deepfake advertisements impersonating a leading diabetes specialist, promoting unproven supplements and discouraging evidence-based treatment – a clear public-health risk.

When people can no longer reliably distinguish authentic information from synthetic content, trust in institutions, expertise and online information degrades – and anxiety follows.

Privacy and surveillance

Privacy has been gradually eroding for years, and this is amplified by AI’s ability to analyse large volumes of personal data – including behavioural, biometric and location data – often without us fully understanding how that data is used.

Pew Research shows persistent public concern about data misuse, impersonation and loss of privacy associated with AI systems.

This anxiety is not limited to government surveillance; it also reflects unease about corporate data practices, profiling, targeted persuasion and information asymmetries between people and institutions.

AI agents ‘taking over’

For most people, fears of an AI apocalypse are not everyday concerns, but they surface during high-profile stories about AI behaving unpredictably or operating autonomously.

One recent example is Moltbook, a platform marketed as a social network for AI agents, which attracted widespread commentary about some weird and disturbing interactions between AI systems.

Reuters reported that the platform had a major security vulnerability that exposed private messages, thousands of email addresses, and over a million credentials, highlighting basic governance and security failures.

These episodes often attract dystopian interpretations reminiscent of science-fiction narratives. But the more immediate risks tend to be practical rather than dramatic: poor security, weak oversight, unclear responsibility and premature deployment.

Concentration of power

Another source of anxiety is the concentration of AI capability among a small number of firms and countries. Many people worry about a future in which a handful of technology giants hold disproportionate power and wealth.

The OECD has noted that generative AI markets may exhibit a ‘winner takes all or most’ dynamic, reinforcing market power and potentially increasing inequality.

When powerful technologies are perceived as unavoidable and foundational, people reasonably ask who benefits, who bears the risk and how accountability is maintained.

Education integrity under pressure

AI anxiety is particularly evident in education.

The concern is not only academic misconduct, but whether assessment continues to measure understanding, reasoning and learning when high-quality outputs can be generated instantly.

A recent article in The Australian reported widespread student use of AI in higher education, including an experiment in which around 80 per cent of 40 student assignments had a high probability of being AI-generated.

This is a global issue.

The United Nations Educational, Scientific and Cultural Organization (UNESCO) has warned that generative AI is advancing faster than institutional readiness, raising concerns about privacy, equity and the long-term future of education.

At its core, this anxiety reflects concern about the purpose of education itself – not just credentialing, but the development of judgment, critical thinking and intellectual independence.

Authenticity and meaning

Finally, there is a quieter concern about authenticity and meaning.

As generative AI becomes capable of producing fluent writing, images and conversation at scale, some people worry less about being replaced and more about being diminished.

They question whether human creativity, effort and connection will continue to be recognised and valued when machine-generated output is ubiquitous.

Research from the Pew Research Center captures this unease.

Many people express concern that AI may reduce human interaction, weaken social connection and devalue human skills and creativity, even while acknowledging its potential benefits.

These concerns are not anti-technology; they reflect a desire for a future in which human contribution remains visible and meaningful alongside increasingly capable machines.

What actually helps if you’re worried about AI

When people raise anxiety with me – whether it’s about their health, the environment or technology – it usually eases once concerns become specific and actionable. AI is no different.

So, here are some tips:

  1. Name the worry: ‘AI’ is too broad to be useful. Are you worried about your job, misinformation, privacy, education or decision-making without oversight?
  2. Clean up your information diet: AI anxiety is often driven by headlines rather than evidence. Limit sensationalist coverage, be cautious with viral screenshots and rely on a small number of trusted sources.
  3. Build your AI literacy: You don’t need to be technical, but you do need to understand how AI is used in your own field – where it helps, where it fails and how outputs should be checked.
  4. Ask for guardrails: Anxiety rises when accountability is unclear. Ask who is responsible when AI is used, how errors are handled and what safeguards exist. Support regulation – even at your workplace or in your home – that focuses on transparency, safety and fairness.

At this moment in history, AI anxiety is not irrational. It reflects rapid change intersecting with our work, education, relationships and identity.

Neither denial nor panic is helpful. Engagement, understanding and our shared responsibility are.

Note: This article was originally published on the University of Melbourne’s Pursuit research news website and republished on DIW with permission. We have been informed that no AI tools were used in creating the text.

Reviewed by Ayaz Khan.

Read next: AI energy use: New tools show which model consumes the most power, and why


by External Contributor via Digital Information World

AI energy use: New tools show which model consumes the most power, and why

Reviewed by Ayaz Khan.

Software can help developers assess AI model energy use, which is necessary to lower costs and strain on the grid

Image: Brett Sayles / Pexels

AI users and developers can now measure the amount of electricity various AI models consume to complete tasks with open-source software and an online leaderboard developed at the University of Michigan.

Companies can download the software to evaluate private models run on private hardware. And while the software can’t evaluate the energy costs of queries run on proprietary AI models at private data centers, it has allowed U-M engineers to measure the power used by open-weight AI models in which the parameters under the hood are publicly available. The power requirements can be viewed on an online leaderboard, which was updated this month. Their results have revealed trends on how AI energy use varies with model design and implementation.

“If you want to optimize energy efficiency and minimize environmental impact, knowing the energy requirements of the models is critical, but popular benchmarks for assessing AI ignore this aspect of performance,” said Mosharaf Chowdhury, associate professor of computer science and engineering and the corresponding author of a study describing the software.

Tools for informed decision-making

The researchers measured energy use across several different tasks, including chatting, video and image generation, problem solving and coding. For some tasks, the energy requirements of open-weight models can vary by a factor of 300. With the results, Chowdhury’s team has developed tutorials for developers to learn how to measure and lower the energy costs of their models. They gave their latest tutorial at the Neural Information Processing Systems (NeurIPS) Conference in December.

The researchers designed their software with partial funding from the National Science Foundation to help solve AI’s growing energy demands. Between 80% and 90% of the sector’s energy is consumed when a trained model processes a request at remote data centers—what the industry calls inference.

As AI models grow in size and are used more often, they need more power. Data centers in the United States consumed about 4% of the country’s total power in 2024—or about as much as Pakistan uses in a year. Data centers are projected to use twice as much power by 2030, according to a study by the Pew Research Center. But many estimates on AI growth rely on ‘envelope’ calculations, which are made by multiplying the maximum power draw per GPU by the number of GPUs. It’s only an estimate of the highest possible energy cost.

“A lot of people are concerned about AI’s growing energy use, which is fair,” Chowdhury said. “However, many who worry can be overly pessimistic, and those who want more data centers are often overly optimistic. The reality is not black and white, and there’s a lot we don’t know because nobody is making direct measurements of AI power use available. Our tool can provide more accurate data for better decision-making.”

Why do some AI models use more power?

The team’s assessments of open-weight models revealed larger trends in how an AI’s design could affect its energy requirements. A key factor was the number of generated tokens—the basic units of data processed by AI. In LLMs, tokens are pieces of words, so wordier models tend to use more energy than concise models. Problem-solving or reasoning models also use more energy because they generate “chains of thought” that contain 10 to 100 more tokens per request.

But the energy requirement of even a single model can change, depending on how it’s run at the data center. Processing queries in batches, for example, will result in less energy use at the data center overall, although larger batches take longer to run. The choice of software for allocating computer memory to queries can also impact AI’s energy requirements.

“There are many ways to deploy AI and translate what the model wants to do into computations on the hardware,” said Jae-Won Chung, U-M doctoral student in computer science and engineering and the study’s first author. “Our tool can automate the search through that parameter space and find the most efficient set of parameters based on the user’s needs.”

The research was also supported by grants and gifts from VMware, the Mozilla Foundation, Cisco, Ford, GitHub, Salesforce, Google and the Kwanjeong Educational Foundation.

Contact: Kate McAlpine.

This post was originally published by the University of Michigan News and is republished here with permission.

Read next: 

AI chatbots provide less-accurate information to vulnerable users, study

• How shaming unethical brands makes companies improve their behaviour
by External Contributor via Digital Information World

How shaming unethical brands makes companies improve their behaviour

Janet Godsell, Loughborough University and Nikolai Kazantsev, University of Cambridge

Recent investigations have uncovered forced labour in agricultural supply chains, illegal fishing feeding supermarket freezers, deforestation embedded in everyday food products, and unsafe conditions in factories producing “sustainable” fashion. These harms were not visible on labels. They surfaced only when journalists, whistleblowers or activists exposed them.

Image: Atoms / Unsplash

And when they did, something predictable happened. Consumers felt uneasy. Brands issued statements. Promises were made. The point is that the force that set change in motion was not regulation. It was consumers.

Discovering that an ordinary purchase may be tied to exploitation or environmental damage creates a jolt of personal responsibility. In our research, we found that when environmental consequences are clearly linked to people’s own buying choices, many are willing to switch products — especially when credible alternatives exist.

But guilt is private. It nudges personal behaviour. It does not automatically reshape systems. The shift happens when private discomfort becomes public voice.

Consumers are often also the first to make hidden environmental harms visible. They post evidence on social media. They question corporate claims. They compare sustainability promises with independent reporting. They organise petitions, boycotts and review campaigns. By shining a spotlight on the truth, the scrutiny shifts from shoppers to brands.

That shift matters because modern brands depend on trust. Reputation is an asset. When sustainability claims are publicly challenged, credibility is at risk. Research in organisational behaviour shows that firms respond quickly to threats to legitimacy. Reputational damage affects customer loyalty, investor confidence and regulatory attention.

In many high-profile cases, supply chain reforms have followed intense public scrutiny rather than quiet compliance checks. Leaders may not act out of moral awakening — but they do act when inaction becomes costly to their reputation.

Consumers can trigger the emotional chain reaction. They feel guilt. They seek information. They speak collectively. That collective voice generates corporate shame.

Sustainability professor Mike Berners-Lee argues in his book A Climate of Truth that demanding honesty is one of the most powerful climate actions available to citizens. Raising standards of truthfulness in business and media changes incentives. When the gap between what companies say and what they do becomes visible, maintaining that gap becomes harder.

Our research explores how that visibility can be strengthened. The findings were clear. When environmental and social consequences are personalised and traceable, sustainability feels less distant. People see both their own role and the role of particular firms. That dual awareness encourages two responses: behavioural change driven by guilt and corporate accountability driven by shame.

Shame works because it is social. Brands care about how they are seen. When the negative environmental and social effects of supply chains can be publicly connected to named products, corporate narratives become contestable in real time.

Making supply chains socially visible

The technology to improve transparency already exists. Companies track goods through logistics systems, supplier databases and digital product-tagging that collect detailed information about sourcing and production. The barrier is not data collection. It is disclosure.

Environmental indicators — carbon emissions, water use, land conversion risk, labour standards compliance — can be linked to products through QR codes or retail apps. Comparable reporting standards would ensure consistency. Simple digital interfaces would make information accessible. Social sharing tools would allow consumers to compare and discuss findings publicly.

Social media is crucial. It already enables workers, communities and campaigners to challenge corporate messaging. Integrating verified supply chain data into these spaces would shift transparency from crisis response to everyday expectation.

This strategy, with its behaviour change directive, could work more effectively than rules or green marketing campaigns alone.

Regulation is essential but often slow and uneven across borders. Marketing campaigns can highlight selective improvements while leaving deeper practices untouched. Transparency activated by collective consumer voice operates differently. It aligns emotional motivation with reputational consequence.

Consumers are not passive recipients of information. They are catalysts. By feeling the first twinge of guilt, asking harder questions and speaking together, they create the conditions under which companies experience shame. When shame threatens trust and market position, change becomes rational and inevitable.

Shame is uncomfortable. But when directed at opaque systems rather than consumers, it can be powerful. By demanding truth and making supply chains socially visible, consumers can push businesses towards greater transparency — and, ultimately, towards more sustainable practice.


Janet Godsell, Dean and Professor of Operations and Supply Chain Strategy, Loughborough Business School, Loughborough University and Nikolai Kazantsev, Postdoctoral Researcher, Institute for Manufacturing, University of Cambridge

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Edited by Asim BN. Reviewed by Ayaz Khan.

Read next: Study: AI chatbots provide less-accurate information to vulnerable users


by External Contributor via Digital Information World