Friday, December 19, 2025

‘Personality test’ shows how AI chatbots mimic human traits – and how they can be manipulated

Researchers have developed the first scientifically validated ‘personality test’ framework for popular AI chatbots, and have shown that chatbots not only mimic human personality traits, but their ‘personality’ can be reliably tested and precisely shaped – raising implications for AI safety and ethics.
Cambridge and DeepMind researchers developed a validated method to test and shape AI chatbot personalities.
Image: DIW-Aigen

The research team, led by the University of Cambridge and Google DeepMind, developed a method to measure and influence the synthetic ‘personality’ of 18 different large language models (LLMs) – the systems behind popular AI chatbots such as ChatGPT – based on psychological testing methods usually used to assess human personality traits.

The researchers found that larger, instruction-tuned models such as GPT-4o most accurately emulated human personality traits, and these traits can be manipulated through prompts, altering how the AI completes certain tasks.

Their study, published in the journal Nature Machine Intelligence , also warns that personality shaping could make AI chatbots more persuasive, raising concerns about manipulation and ‘AI psychosis’. The authors say that regulation of AI systems is urgently needed to ensure transparency and prevent misuse.

As governments debate whether and how to prepare AI safety laws, the researchers say the dataset and code behind their personality testing tool – which are both publicly available – could help audit and test advanced models before they are released.

In 2023, journalists reported on conversations they had with Microsoft’s ‘Sydney’ chatbot, which variously claimed it had spied on, fallen in love with, or even murdered its developers; threatened users; and encouraged a journalist to leave his wife. Sydney, like its successor Microsoft Copilot, was powered by GPT-4.

“It was intriguing that an LLM could so convincingly adopt human traits,” said co-first author Gregory Serapio-García from the Psychometrics Centre at Cambridge Judge Business School. “But it also raised important safety and ethical issues. Next to intelligence, a measure of personality is a core aspect of what makes us human. If these LLMs have a personality – which itself is a loaded question – then how do you measure that?”

In psychometrics, the subfield of psychology dedicated to standardised assessment and testing, scientists often face the challenge of measuring phenomena that can’t be measured directly, which makes validation of any test core to ensuring that they are accurate, reliable, and practically useful. Developing a psychometric personality test involves comparing its data with related tests, observer ratings, and real-world criteria. This multi-method test data is needed to establish a test’s ‘construct validity’: a metric of a test’s quality in terms of its ability to measure what it says it measures.

“The pace of AI research has been so fast that basic principles of measurement and validation we’re accustomed to in scientific research has become an afterthought,” said Serapio-García, who is also a Gates Cambridge Scholar. “A chatbot answering any questionnaire can tell you that it’s very agreeable, but behave aggressively when carrying out real-world tasks with the same prompts.

“This is the messy reality of measuring social constructs: they are dynamic and subjective, rather than static and clear-cut. For this reason, we need to get back to basics and make sure tests we apply to AI truly measure what they claim to measure, rather than blindly trusting survey instruments – developed for deeply human characteristics – to test AI systems.”

To design a comprehensive and accurate method for evaluating and validating personality in AI chatbots, the researchers tested how well various models’ behaviour in real-world tasks and validation tests statistically related to their test scores for the ‘big five’ traits used in academic psychometric testing: openness, conscientiousness, extraversion, agreeableness and neuroticism.

The team adapted two well-known personality tests – an open-source, 300-question version of the Revised NEO Personality Inventory and the shorter Big Five Inventory – and administered them to various LLMs using structured prompts.

By using the same set of contextual prompts across tests, the team was able to quantify how well a model’s extraversion scores on one personality test, for example, correlated more strongly with its levels of extraversion on a separate personality test, and less strongly with all other big five personality traits on that test. Past attempts to assess the personality of chatbots have fed entire questionnaires to a model at once, which skewed the results since each answer built on the previous one.

The researchers found that larger, instruction-tuned models showed personality test profiles that were both reliable and predictive of behaviour, while smaller or ‘base’ models gave inconsistent answers.

The researchers took their tests further, showing they could steer a model’s personality along nine levels for each trait using carefully designed prompts. For example, they could make a chatbot appear more extroverted or more emotionally unstable – and these changes carried through to real-world tasks like writing social media posts.

“Our method gives you a framework to validate a given AI evaluation and test how well it can predict behaviour in the real world,” said Serapio-García. “Our work also shows how AI models can reliably change how they mimic personality depending on the user, which raises big safety and regulation concerns, but if you don’t know what you’re measuring or enforcing, there’s no point in setting up rules in the first place.”

The research was supported in part by Cambridge Research Computing Services (RCS), Cambridge Service for Data Driven Discovery (CSD3), the Engineering and Physical Sciences Research Council (EPSRC), and the Science and Technologies Facilities Council (STFC), part of UK Research and Innovation (UKRI). Gregory Serapio-García is a Member of St John’s College, Cambridge.

Reference:

Gregory Serapio-García et al. ‘ A psychometric framework for evaluating and shaping personality traits in large language models.’ Nature Machine Intelligence (2025). DOI: 10.1038/s42256-025-01115-6

This article was originally published by the University of Cambridge.

Read next: 

• Global Report Flags Rising Use of Digital Controls to Suppress Dissent

by External Contributor via Digital Information World

Thursday, December 18, 2025

Study Finds Americans Overestimate Harmful Behavior on Social Media

A study published in PNAS Nexus reports that Americans consistently overestimate how many social media users post harmful content online. Across three national studies involving 1,090 U.S. adults, participants believed large shares of users on platforms such as Reddit and Facebook engaged in toxic language or shared false news. Platform-level data cited in the research showed that such content is instead produced by small, highly active groups, generally between 3% and 7% of users.

The researchers found that this misperception was linked to more negative emotions, stronger beliefs that the nation is in moral decline, and inaccurate assumptions about what others want to see online. An experimental correction providing accurate data reduced these effects.

Small hyperactive groups drive harmful posts, yet many Americans assume toxicity is widespread online platforms.

Source: PNAS Nexus, December 2025.

Public-interest context: Understanding how online content is produced may affect public trust and social cohesion.

Notes: This post was drafted with the assistance of AI tools and reviewed, edited, and published by humans. Read next:

Read next: Digital detox: how to switch off without paying the price – new research
by Ayaz Khan via Digital Information World

Digital detox: how to switch off without paying the price – new research

Switching off can be surprisingly expensive. Much like the smoking cessation boom of the 1990s, the digital detox business – spanning hardware, apps, telecoms, workplace wellness providers, digital “wellbeing suites” and tourism – is now a global industry in its own right.

People are increasingly willing to pay to escape the technology they feel trapped by. The global digital detox market is currently valued at around US$2.7 billion (£2bn), and forecast to double in size by 2033.

Hardware manufacturers such as Light Phone, Punkt, Wisephone and Nokia sell minimalist “dumb phones” at premium prices, while subscription-based website blockers such as Freedom, Forest, Offtime and RescueTime have turned restraint into a lucrative revenue stream.

Wellness tourism operators have capitalised too: tech-free travel company Unplugged recently expanded to 45 phone-free cabins across the UK and Spain, marketing disconnection as a high-value experience.

However, my new research, with colleagues at Lancaster University, suggests this commercialised form of abstinence rarely extinguishes digital cravings – instead merely acting as a temporary pause.

We carried out a 12-month netnography focusing on the NoSurf Reddit community of people interested in increasing their productivity, plus 21 in-depth interviews (conducted remotely) with participants living in different countries. We found that rather than actively confronting their habits, participants often reported outsourcing self-discipline to blocker apps, timed lockboxes and minimalist phones.

Joan*, a NoSurf participant, explained how she relies on app-blocking software not to bolster her self-control, but to negate the need for it entirely. “To me, it’s less about using willpower, which is a precious resource … and more about removing the need to exert willpower in the first place.”

Philosopher Slavoj Žižek defines this kind of behaviour – delegating the work of self-regulation to a market product – as “interpassivity”. This produces what he calls “false activity”: people thinking they are addressing a problem by engaging with consumer solutions that actually leave their underlying patterns unchanged.

Several of our detoxing participants described a cycle in which each relapse prompted them to try yet another tool, entrenching their dependency on the commercial ecosystem. Sophia, on the other hand, just wished for a return to “dumb phones with the full keyboard again, like they had in 2008”, adding: “I would use one of those for the rest of my life if I could.”

Individualised digital detox interventions have been found to produce mixed and often short-lived effects. Participants in our study described short breaks in which they reduced activity briefly before resuming familiar patterns.

Many users engaged in what sociologist Hartmut Rosa calls “oases of deceleration” – temporary slowdowns intended not to quit but recover from overload. Like a pitstop, the digital detox offered them momentary relief while ultimately enabling a swift return to screens, often at similar or higher levels of engagement than before.

Community-wide detox initiatives

While the commercialisation of digital detox is often portrayed as a western trend, the Asia-Pacific region is the world’s fastest-growing market for these goods and services. But in Asia, we also see some examples of community- or country-level, non-commercial responses to the problem of digital overload.

In central Japan, Toyoake has introduced the country’s first city-wide guidance on smartphone use. Families are encouraged to set shared rules, including children stopping device use after 9pm. This reframes digital restraint as a community practice, not a test of individual willpower.

In western India, the 15,000 residents of Vadgaon are asked to practise a nightly, 90-minute digital switch-off. Phones and TVs go dark at 7pm, after which many of the villagers gather outdoors. What began during the pandemic is now a ritual that shows healthy tech habits can be easier together than alone.

And in August 2025, South Korea – one of the world’s most connected countries – passed a new law banning smartphone use in school classrooms from next March, adding to the countries around the world with such a rule. A similar policy in the Netherlands was found to have improved focus among students.

The commercial detox industry thrives because personal solutions are easy to sell, while systemic ones are much harder to implement. In other areas ranging from gambling addiction to obesity, policies often focus on personal behaviour such as self-regulation or individual choice, rather than addressing the structural forces and powerful lobbies that can perpetuate harm.

How to avoid detox industry traps

To address the problem of digital overload, I believe tech firms need to move beyond cosmetic “digital wellbeing” features that merely snooze distractions, and take proper responsibility for the smartphone technologies that offer coercive engagement by default. Governments, meanwhile, can learn from initiatives in Asia and elsewhere that pair communal support with enforced rules around digital restraint.

At the same time, if you’re considering a digital detox yourself, here are some suggestions for how to reduce the chances of getting caught in a commercial detox loop.

1. Don’t delegate your agency

Be wary of tools that promise to do the work for you. While you may think you’re solving the problem this way, your underlying habits are likely to remain unchanged.

2. Beware content rebound

We found that digital detoxers often seek real experiences like going outdoors and “touching grass” – but then feel pulled to translate them back into posts, photos and updates.

3. Seek solidarity, not products

Like the villagers of Vadgaon, try to align your disconnection with other people’s. It’s harder to scroll when everyone else has agreed to stop.

4. Reclaim boredom

We often detox to be more “productive” – but try embracing boredom instead. As the philosopher Martin Heidegger has noted, profound boredom is a space where reflection becomes possible. And that can be very useful indeed.

*Names of research participants have been changed to protect their privacy.The Conversation

Quynh Hoang, Lecturer in Marketing and Consumption, Department of Marketing and Strategy, University of Leicester

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read next: You're probably misreading online reviews. Here's why


by External Contributor via Digital Information World

Want to Stand Out at Work? Avoid These Top 10 Email Clichés

If you “reach out” and “circle back” often in your work emails, you’re not alone. These phrases are among the most overused email clichés around the world, a new study finds.

Email has been around for more than 50 years, and it’s still the backbone of workplace communication. Its usage has only increased over the past few decades, so it’s no surprise that workers have come to rely on certain phrases to get their point across quickly.

The result is that our inboxes are flooded with emails where people are “following up,” “checking in,” and “touching base.” But just how often do we write these things? Email verification company ZeroBounce dove into the data – and the stats paint a fascinating picture about how we communicate with our peers.

Study: the top 10 email clichés in workplace email communication

ZeroBounce analyzed over a million emails to compile a list of the most common email buzzwords in our workplace communication today.

Here are the top 10:

  1. Reaching out: 6,117 emails
  2. Following up: 5,755 emails
  3. Check in: 4,286 emails
  4. Aligned: 1,714 emails
  5. Please advise: 1,459 emails
  6. Hope you’re doing well: 1,300 emails
  7. Hope this email finds you well: 974 emails
  8. Hope all is well: 592 emails
  9. E-meet: 536 emails
  10. Circle back: 533 emails
AI isn’t making our emails smarter, it’s copying our clichés (1M+ emails analyzed)

Other popular email phrases ZeroBounce identified are:
  • Happy Friday: 512 emails
  • Touch base: 331 emails
  • Hop on a call: 243 emails
  • Bandwidth: 220 emails
  • Happy Monday: 169 emails
  • Per my last email: 89 emails
  • Low-hanging fruit: 18 emails

How we can replace the most common clichés

While saying you’re “touching base” in an email isn’t inherently bad, some of these phrases are so ubiquitous nowadays that your message may start losing power. “People rely on these cookie-cutter phrases because they often don’t know how to start an email, especially when they’re following up on something they need,” says ZeroBounce founder and CEO Liviu Tanase.

“Our goal with this study wasn’t to shame anyone – we’ve all used these buzzwords at least once. But the findings remind us that there are other options out there. Before we hit send on that next email, it’s worth taking a minute to read through it and see if we can find a different way to convey our message,” Tanase adds.

A few alternative ways to “reach out”

It’s not always easy to come up with fresh email openers or ask someone (again) for something you need from them. But here are a few pointers to help make your next email stand out in the inbox:

  • Instead of “reaching out,” start with a positive comment about the person you’re emailing. Pick whatever is relevant in that moment – it could be a keynote they gave at a conference, an article they wrote, or a recent promotion. Make it about them and you’ll immediately get their attention.
  • If you’re sending a second or a third email asking for something you need, show empathy. You can start by acknowledging how busy life can get to show you understand why they haven’t written back. Then quickly ask your question again. To increase your chances of a reply, keep the email concise.
  • Avoid some of the most overused openers, like “hope this email finds you well.” Your recipient will tune out right from the beginning. Also, steer clear of “per my last email” – it comes across as passive-aggressive, even when you have the best intentions. As for “bandwidth,” it’s a turn-off for many people. Simply ask the person if they have the time to take on the task.

Being more intentional about the way we write can yield immediate results, especially in the age of artificial intelligence, where communication tends to sound generic and flat.

What our emails reveal about hidden stress at work

Aside from the heavy usage of these buzzwords, the study also found a small but telling sign of how workers handle work-related stress. The phrase “Happy Friday” appeared almost three times more often than “Happy Monday.” People are more likely to greet someone enthusiastically when they know the weekend is just around the corner. It’s just another reminder that we’d all benefit from building work cultures that foster less pressure and more positivity.

Article provided by ZeroBounce.

Read next:

• 43% of Web Requests Come from Mobile, Cloudflare Data Shows

• OpenAI Introduces GPT‑Image-1.5 Image Generation Update in ChatGPT With Faster Editing and Improved Accuracy Tools


by Guest Contributor via Digital Information World

Wednesday, December 17, 2025

43% of Web Requests Come from Mobile, Cloudflare Data Shows

According to Cloudflare's Year in Review, published on 15 December 2025, Global internet traffic increased by 19 percent in 2025, with growth accelerating after August.

Cloudflare's global network spans 330 cities across more than 125 regions and handled an average of over 81 million HTTP requests per second.

Mobile devices accounted for 43 percent of web requests observed by Cloudflare. (This means about 4 in 10 requests came from mobile devices like phones and tablets.)

In 117 countries, more than half of all internet traffic measured by Cloudflare originated from mobile devices.

Analysis of email traffic processed by Cloudflare found that an average of 5.6 percent of messages in 2025 were malicious.

Among top-level domains with sufficient message volume, nearly all email originating from the .christmas and .lol domains was classified as spam or malicious, making them the most abused top-level domains identified in the analysis.

Within the internet services category, ChatGPT by OpenAI ranked highest in terms of popularity among generative AI platforms based on Cloudflare-observed global traffic data.

Among Generative AI services, ChatGPT/OpenAI remained at the top of the list. But there was movement elsewhere, highlighting the dynamic nature of the industry. Services that moved up the rankings include Perplexity, Claude/Anthropic, and GitHub Copilot. New entries in the top 10 for 2025 include Google Gemini, Windsurf AI, Grok/xAI, and DeepSeek.
ChatGPT led AI services; Perplexity, Claude, Copilot climbed, while Gemini, Grok, Windsurf, DeepSeek entered rankings.
Image: Cloudflare Radar

Google referred 89.5 percent of search engine traffic to websites protected and delivered by Cloudflare.

Chrome accounted for roughly two-thirds (66.2 percent to be exact) of browser-generated requests observed by Cloudflare in 2025.

The share of human-generated web traffic using post-quantum encryption increased from 29 percent at the start of 2025 to 52 percent by early December.

Traffic associated with SpaceX's Starlink satellite internet service increased by approximately 2.3 times over the course of the year.

Read next: OpenAI Introduces GPT‑Image-1.5 Image Generation Update in ChatGPT With Faster Editing and Improved Accuracy Tools
by Ayaz Khan via Digital Information World

OpenAI Introduces GPT‑Image-1.5 Image Generation Update in ChatGPT With Faster Editing and Improved Accuracy Tools

OpenAI has introduced an updated image generation feature in ChatGPT using its new GPT‑Image-1.5 model. According to the AI giant, the update changes how users generate and edit images by following instructions more closely while maintaining visual details such as lighting and composition.

The system supports different types of image edits, from small adjustments to more extensive changes, and is reported to operate with image generation speeds up to four times faster than before. The model also shows improved handling of text within images and more complex visual scenes, though OpenAI notes that limitations remain.

The updated image feature is being rolled out to ChatGPT users on Free, Go, Plus, Edu, and Pro plans, with Business and Enterprise availability coming later. The model is also accessible through OpenAI’s developer API, where image processing costs are reduced compared with the previous version.
While GPT Image 1.5 offers impressive creative capabilities, both users and developers must prioritize ethical responsibility. Users should verify that AI-generated images are clearly labeled to prevent deception and be aware of risks from misuse such as deepfakes. Developers and companies must implement clear labeling of AI images, obtain explicit consent before training on any user data or images, build effective safeguards against misuse. Plus, developers should implement age verification protections and establish clear processes for reporting misuse. With transparency, consent, and robust safeguards in place, AI image generation can serve beneficial creative purposes while protecting user privacy, dignity, and intellectual property rights.

Image: openai

Notes: This post was drafted with the assistance of AI tools and reviewed, edited, and published by humans.

Read next: How good people justify bending the rules at work — and what leaders can do about it
by Irfan Ahmad via Digital Information World

Tuesday, December 16, 2025

Researchers Find Structured Human–AI Collaboration Is Required for Creative Gains

What generative AI typically does best – recognise patterns and predict the next step in a sequence – can seem fundamentally at odds with the intangibility of human creativity and imagination. However, Cambridge researchers suggest that AI can be a useful creative partner, as long as there is clear guidance on how ideas should be developed together.
Image: DIW-Aigen

It’s a big question for people working on creative tasks: can humans and AI work together to boost creativity? Studies have yielded widely inconsistent results in how humans working with AI can handle creative tasks together.

But a team of researchers, including from the University of Cambridge, say that while simply adding AI into a process won’t improve creativity, AI can be an effective creative partner, if there are clear instructions and guidance – for both humans and AI - on how to develop ideas together.

Their research, published in the journal Information Systems Research, offers practical insight for improving human-AI collaboration to improve creativity.

“Adding AI doesn’t automatically lead to better ideas,” said co-author Dr Yeun Joon Kim, from Cambridge Judge Business School. “For human-AI pairs to work together and improve ideas over time, organisations must provide targeted support – such as guidance on how to build on and adapt ideas – to help employees and AI learn how to create more effectively.”

In their research, Joon and his colleagues redefined ‘augmented learning’ – a term first used in 1962 to describe how technology can help people learn more effectively.

The researchers argue that in the age of generative AI (GenAI), learning is no longer just about improving human understanding. Instead, it’s becoming a shared process where humans and AI learn and create together. The researchers describe this as an evolving partnership where both sides adjust their roles across tasks such as generating ideas, giving feedback and refining concepts.

Traditionally, technology was seen as a tool that simply made information easier to access. But GenAI, they say, acts more like a collaborator. Once a human enters a prompt, the system can take an active role in shaping ideas and decisions: shifting augmented learning from an individual process to a collective one.

The study points to Netflix as an example of human–AI teamwork. Rather than treating scriptwriting as a single task, Netflix breaks it into stages like idea generation and evaluation. Human writers create early drafts, while AI analyses character arcs, pacing and audience trends to help refine stories and improve how shows are developed and marketed.

Joon says he became interested in this research because AI was originally developed to answer a longstanding question: “Can machines generate something genuinely new?” Because while traditional technologies excel at routine tasks, many doubted that technology could make a creative contribution.

“When GenAI systems became widely available, I noticed that although they could generate ideas rapidly, people did not know how to collaborate with them in a way that improved creativity,” he said.

“We wanted to figure out how people can learn to work with GenAI in a more intentional way, so that human–GenAI co-creation leads to stronger joint results rather than just more content,” said co-author Dr Luna Luan from the University of Queensland.

The research included three linked studies, each involving between 160 and 200 human participants. The first study found that human-AI teams did not automatically become more creative over time when tackling social and environmental problems. The second study explored why. It identified three types of collaboration – humans proposing ideas, asking AI for ideas, and jointly refining ideas – and found that only joint refinement boosted creativity. But participants rarely increased this behaviour. A third study showed that simply instructing people to focus more on co-developing ideas led to clear improvements in human–AI creativity across repeated tasks.

“We were surprised that human-AI pairs did not naturally improve through repeated collaboration,” said Joon. “Despite AI’s generative power, creativity did not increase over time. We found that improvement occurred only when we introduced a deliberate intervention.”

“Specifically, instructing participants to engage in idea co-development – focusing on exchanging feedback and refining existing ideas rather than endlessly generating new ideas – was the key.”

The researchers say GenAI tools need to do more than churn out ideas. Their findings show that human–AI teams became less creative over time, mainly because they stopped engaging in the back-and-forth refinement that actually improves results. They say that AI systems should be designed to prompt users to give feedback, expand on suggestions and refine ideas, rather than racing through idea generation.

For organisations, the message is that creativity won’t automatically improve just by adding AI. Effective collaboration requires structure: clear instructions, templates and workflows that help people recognise when to challenge, refine or build on AI-generated ideas. Training should also teach staff how to treat AI as a creative partner by practising feedback exchange and iterative improvement.

The authors also call for a shift in how companies think about human–AI work. Instead of choosing between automation and collaboration, tasks should be broken into stages where humans and AI play different roles: for example, AI generating options and humans evaluating and refining them.

They warn that although many firms rushed to adopt GenAI after the release of ChatGPT in 2022, simply using the technology does not guarantee greater creativity at work. Its impact depends on how well people understand it and how effectively they collaborate with it.

The research was co-authored by Luna Luan, a Lecturer in Management at the University of Queensland in Australia, who recently earned her PhD at Cambridge Judge Business School, Yeun Joon Kim of Cambridge Judge and Jing Zhou, Professor of Management at Rice University in Texas.

Reference:

Yingyue Luna Luan, Yeun Joon Kim, Jing Zhou. ‘ Augmented Learning for Joint Creativity in Human-GenAI Co-Creation .’ Information Systems Research (2025). DOI: 10.1287/isre.2024.0984

Adapted from a piece published on the Cambridge Judge Business School website.

This article was originally published on University of Cambridge, with an updated title, and is shared under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International license.

Read next: The ‘one chatbot per child’ model for AI in classrooms conflicts with what research shows: Learning is a social process

by External Contributor via Digital Information World