Monday, September 8, 2025

Study Finds Human-Written Crisis Messages Viewed as More Credible Than AI

Artificial intelligence is increasingly used in corporate communication, but new research shows it may not be suited for sensitive situations. A study in Corporate Communications: An International Journal found that crisis responses attributed to people were judged as more credible and more helpful to a company’s reputation than identical messages said to come from AI.

Testing Trust in Crisis Responses

Researchers built an experiment around a fictional company called Chunky Chocolate, which was described as facing backlash after reports that its products made customers sick. Participants read one of six possible press releases. Each message had the same content but differed in two ways: whether it was written by a person or by AI, and whether the tone was informational, sympathetic, or apologetic.

The study involved 447 students in journalism and communication programs at a Midwestern university. They evaluated the credibility of the message, the credibility of its source, and the company’s reputation after reading the release.

Human Messages Scored Higher

Results showed a clear pattern. Messages labeled as human-written were rated higher across all measures. On a seven-point scale, human sources received an average credibility score of 4.40, compared with 4.11 for AI. For message credibility, human versions averaged 4.82 while AI versions scored 4.38. Company reputation followed the same trend, with averages of 4.84 for human messages and 4.49 for AI.

Because the content of the statements was unchanged, the difference came only from how authorship was presented. Labeling a release as AI-generated lowered trust, even when the words were identical.

Tone Had Little Effect

Researchers expected an apologetic or sympathetic tone to influence perceptions. Participants did notice the different tones, but ratings of credibility and reputation did not vary much. The communicator’s identity carried more weight than the style of the message.

What It Means for Public Relations

AI already plays a role in public relations through tasks like media monitoring, content targeting, and social media management. Some suggest using it to draft press releases or respond to crises. The study points to risks in doing so, since audiences seem less likely to trust a crisis message when it is tied to AI.

Limits of the Study

The experiment used a fictional company and a student sample, which may not represent the wider public. Participants’ familiarity with digital tools and AI could also shape their views. Another factor is the explicit labeling of AI authorship, as real companies may not always disclose when AI is used.

Even with these limits, the research indicates that audiences still place greater trust in human credibility during moments of public scrutiny.


Notes: This post was edited/created using GenAI tools. Image: DIW-Aigen.

Read next:

• Artificial Sweeteners Linked to Faster Memory Decline in Midlife

• Study Finds LLM Referrals Convert At 4.87% Versus 4.6% For Search, But Scale Remains Tiny
by Irfan Ahmad via Digital Information World

Sunday, September 7, 2025

Study Finds LLM Referrals Convert At 4.87% Versus 4.6% For Search, But Scale Remains Tiny

A six-month analysis of 54 websites found that traffic from large language models converts at almost the same rate as organic search. The research, carried out by Amsive, used Google Analytics 4 data from sites with validated purchases or form fills.

Conversion Rates

Organic visits converted at 4.6 percent. LLM referrals came in at 4.87 percent. On the surface that looked like a gain for LLMs, but statistical testing showed the difference was not significant. In other words, both channels brought in users who converted at nearly identical rates.

Site-Level Differences

Results varied from one website to another. Some saw LLM referrals converting better than their averages, others saw weaker performance. Just over half of the sample leaned positive for LLM, but not by a wide margin. The split highlighted how much outcomes depend on how AI tools select and surface content.

Higher-Volume Sites

Filtering for larger websites, those with at least 100,000 sessions and enough LLM traffic to test, produced a bigger gap: organic at 5.81 percent and LLM at 7.05 percent. Even then the edge failed to clear statistical tests. The analysis showed that the apparent lift could be explained by random variation.

Business Models


Breaking the data into B2B and B2C websites did not change the picture. B2B sites converted at 2.03 percent from LLM referrals and 1.68 percent from organic. B2C sites converted at 10.31 percent from LLM and 8.50 percent from organic. Neither difference was large enough to be reliable once tested for significance.

Industry Patterns

By industry, outcomes were mixed. Financial services and travel recorded higher conversion rates from LLMs. E-commerce and consumer services leaned toward organic. Because sample sizes in each vertical were small, no firm conclusion could be drawn.

Traffic Share

The study found scale to be the critical factor. LLM referrals accounted for less than one percent of total sessions across the dataset. Organic search made up nearly a third of all visits and conversions. In fact, about nine out of ten websites saw LLM traffic contribute less than 0.6 percent of sessions.

Study Limits

The research measured only macro conversions such as purchases and lead forms. It did not track how many leads became paying customers. Conversions were counted on a session basis and attributed to the last click, so earlier touchpoints were not included.

Key Takeaway

For now, organic search remains the leading channel for both scale and consistency. LLM referrals may grow in importance as usage expands, but current evidence shows they are not outperforming search in conversion terms. Businesses may want to monitor LLM traffic closely while continuing to treat search as the foundation of their digital strategy.

 Notes: This post was edited/created using GenAI tools.

Read next:

• Bad Sign-Up Flows Cost SaaS Companies Customers Before They See the Product

• How Many Prompts Can You Run on Gemini Each Day? Google Finally Sets the Numbers
by Irfan Ahmad via Digital Information World

Bad Sign-Up Flows Cost SaaS Companies Customers Before They See the Product

With SaaS, you've got no second chance to make an entrance. Customers come through the door, sign up, and demand to be given instant access. Anything less and they disappear forever. A new Frontegg survey verifies just how much you stand to lose: 15% of SaaS users never come back after a less-than-perfect login experience. Frontegg's First Login Benchmark Report surveyed 439 SaaS users to find out what they crave most from onboarding and why companies get it fundamentally wrong. What their findings reveal is just how little SaaS products need to do in order to earn user trust and, having lost it, how rarely the window of opportunity reopens.



The First Impressions Aren't Skin-Deep Quiz

Onboarding is not a handshake anymore. It's a background check, trust evaluation, and product demo all in one. Nearly half of the users (48%) reported that they had abandoned a SaaS service due to registration taking too long. They would have already decided to return or not by the time they reached the dashboard. 36% churned on confirmation by e-mail step, and 21% insisted on instant access on sign-up. This go-fast strategy isn't one taken in an attempt to compromise on security at all. It truly was the case, as speed and security did factor into 46% of those making their initial impression of a SaaS solution. Just 32% prioritized security over speed. That fast and stable combination is hard to find and nearly impossible to regain.

Why Login UX Can Make or Break Trust

Over half the users (58%) would not even subscribe to a SaaS application whose logon process did not look above board. That's a harsh term, but it's the best word to use for the gut-level feeling many users get when design, language, or interaction flow triggers a sense of unease. It isn't rational, but perceptions govern. The majority of the irritations were old favorites. CAPTCHA was the most common top-of-the-list trust-breaking authentication process, seconded by convoluted multi-step identification verifications and cumbersome email verifications. The processes may lead to added security, but are most typical indicators of lack of design maturity or cumbersome flow if done ineptly. And the psychological impact is real. Twenty-one percent of the surveyed users admitted to rage-quitting onboarding by bailing on the entire process right there out of frustration long before even getting to experience the product. Those numbers show a bigger pattern. Login UX was once ambient. It then turned into the product experience itself.

The Long-Term Relevance of Login Interfaces

The most secure login features, most strongly linked in users' minds with safety and professionalism, come as no surprise yet demonstrate a persistent gap between desired and delivered.

Two-factor authentication (2FA): 74% of users reported feeling safer when it's enabled

Email confirmation: 23% consider it a sign of trust if executed well

Single sign-on and social login: 20% and 11%, respectively, value the convenience of them

These statistics indicate why intelligent design and simplicity are key when it comes to establishing user confidence. Users do not want the verification to be omitted completely, but rather quick, secure, and transparent.

Cost of a Bad Experience per Person

SaaS products have only one chance. 53% of the users would return to a SaaS product after a bad login experience, according to Frontegg, yet 15% would never return. Even if they do remain, they use it fractionally. Within a 30-day trial, over half (51%) had used 25 to 50% of the product features. That extent of selective adoption does more than limit the initial impression. It holds product stickiness and long-term satisfaction at bay. It is not just first-week churn. The negative sign-up impact extends into usage, word of mouth, and upsells.

What SaaS Teams Can Do Differently and Learn from

Understanding where users drop off is a crucial first step, but the tougher part is determining how to remove the friction causing the drop off, without implementing something that adds friction.

Here are some areas of focus for SaaS teams creating a login flow or reimagining one:

Removing friction without decreasing security. A fast way to access your app is important, but fast access will be no good and will not build trust if it amounts to a sketchy login. Users are willing to accept 2FA if it works properly; ineffective versions of Captcha or multi-step verification will likely drive users in the opposite direction.

Making verification frictionless. Email verification is a common drop point; real-time verification messages or progressive disclosure might alleviate friction at this delicate moment of onboarding for your users.

Treating the login UX process like the design of the core product. All screens and steps should feel purposeful for users. Confusion over where to go next or confusion over what the progress appears as can damage trust faster than slow-loading issues.

Testing sign-up as a potential new user. Internal teams often gloss over and ignore obstacles that new users frequently encounter. Testing, along with real user feedback, will provide the insights that metrics can miss.

The Big Picture: Trust Before Features

SaaS functionality and workflow take months or years to build. It will be useless if the users never even go beyond the front door. Frontegg data exemplifies the cultural pivot. Onboarding isn't just getting users past the front door. It's the first promise of value, security, and empathy. And when virtually one in every five users rage-quits on the first day and an additional 15% ghost after the one bum login, the margin of error basically does not exist anymore. For SaaS companies, that's it. Good onboarding is not a nice-to-do. Growth. Retention. Reputation. Done. The virtual handshake is now. SaaS companies that wish to retain customers long enough to show them what they're providing must trust customers immediately and often. That starts on the login page. The full First Login Benchmark Report results reveal more about why and when customers churn SaaS products prior to use. Get the full study here to read the recommendations and benchmarks.

Read next: How Many Prompts Can You Run on Gemini Each Day? Google Finally Sets the Numbers


by Irfan Ahmad via Digital Information World

Why AI Chatbots Still Hallucinate: Researchers Trace Errors to Training Data Gaps and Misaligned Benchmarks

Artificial intelligence tools are now used in classrooms, offices, and customer support desks. Yet they carry a flaw that refuses to fade. Ask a chatbot a simple factual question, and it may deliver a confident answer that turns out to be wrong. Researchers at OpenAI, joined by collaborators at Georgia Tech, say they now have a clearer picture of why this happens.

Where the Mistakes Begin

Large language models are trained by scanning enormous volumes of text and learning to predict what word should come next. That process gives them fluency, but it also builds in errors. The team’s paper explains that even with perfectly clean training data, mistakes are mathematically inevitable.

Some facts are simply too rare for a system to learn. A birthday that appears once in a dataset, for example, provides no pattern the model can generalize. The authors call this the “singleton rate.” High singleton rates mean a model will almost certainly invent details when asked about them. This is why common knowledge tends to be correct, while obscure details often come back scrambled.

From Exams to Algorithms

The training phase is only half the story. After that, models are fine-tuned to better match human expectations. But the way they are tested keeps the cycle going.

Benchmarks usually grade answers as right or wrong. There’s no credit for admitting uncertainty. A chatbot that says “I don’t know” is punished as harshly as one that blurts out something false. Under that system, guessing is the smarter move. Over time, models are effectively trained to bluff.

The researchers compare this to multiple-choice exams. Students who leave blanks score lower than those who make lucky guesses. AI models, shaped by similar scoring, act in much the same way.

When Models Go Wrong

Examples from the study illustrate how deep the problem runs. One widely used model was asked for Adam Kalai’s birthday — Kalai being one of the paper’s authors. It gave three different dates across separate attempts. None were right, and it had been told to answer only if certain.

In another test, a system failed at counting the letters in a word, producing results that made little sense. These cases show both the arbitrary fact problem and what the authors call poor model representation, where the structure of the system limits its ability to handle simple tasks.

Changing the Scoreboards

The researchers argue the solution lies in evaluation. Instead of rewarding risky guesses, new benchmarks should penalize confident wrong answers more than admissions of uncertainty. One option is to grant partial credit when a model holds back. Another is to set confidence thresholds in the test instructions, telling the model to answer only if it reaches a defined level of certainty.

This echoes older exam systems where wrong guesses were penalized, discouraging blind attempts. The same principle could shift AI development toward models that value accuracy over bravado.

Limits and Outlook

The study makes clear that hallucinations will not vanish completely. Some questions are inherently unanswerable because the data is missing, ambiguous, or too complex. But better testing could reduce the most damaging errors and build greater trust in AI systems.

The broader point is that hallucinations are not random glitches. They are the product of how models are trained, and more importantly, how they are judged. If the industry changes the scoreboards, the behavior of the models is likely to follow.

Notes: This post was edited/created using GenAI tools. 

Read next:

• AI Models Can Now Run Ransomware Attacks on Their Own, Study Finds

• Secure Online Transactions and Business Models in E-commerce and Marketplaces

• Chatbots Are Spreading More False Claims, NewsGuard Report Shows


by Irfan Ahmad via Digital Information World

Secure Online Transactions and Business Models in E-commerce and Marketplaces

This article was created in partnership with Mangopay for promotional purposes.

In the digital age, the landscape of commerce has dramatically shifted from traditional brick-and-mortar stores to online platforms. This transformation has given rise to two dominant business models: e-commerce and marketplaces. Both models have revolutionized the way consumers shop and businesses operate, offering unparalleled convenience and access to a global market. However, with these advancements come challenges, particularly in ensuring secure online transactions. As cyber threats become more sophisticated, the need for robust security measures in e-commerce and marketplaces is more critical than ever.


This article delves into the intricacies of these business models, explores the differences between them, and discusses strategies for enhancing transaction security to foster trust and reliability in the digital marketplace.

Understanding the Difference Between E-commerce and Marketplaces

E-commerce and marketplaces are often used interchangeably, but they represent distinct business models with unique characteristics. E-commerce refers to the buying and selling of goods and services over the internet. Typically, e-commerce platforms are operated by a single vendor who manages the entire sales process, from product listing to payment processing and delivery. This model allows businesses to maintain control over their brand and customer experience.

On the other hand, marketplaces are platforms that connect multiple sellers with buyers. These platforms do not own the inventory but facilitate transactions between third-party vendors and consumers. Marketplaces offer a wide variety of products from different sellers, providing consumers with diverse options and competitive pricing. Examples of popular marketplaces include Amazon, eBay, and Etsy.

For a more detailed exploration of the differences between these two models, you can visit https://blog.mangopay.com/en/home/what-is-the-difference-between-e-commerce-and-marketplaces.

Enhancing Payment Security in E-commerce and Marketplaces

As online transactions become more prevalent, ensuring payment security is paramount for both e-commerce platforms and marketplaces. Consumers need assurance that their financial information is protected from fraud and unauthorized access. To achieve this, businesses must implement robust security measures, including encryption, tokenization, and secure payment gateways.

Encryption is a fundamental security measure that protects sensitive data by converting it into a code that can only be deciphered with a key. Tokenization replaces sensitive data with unique identifiers, or tokens, that have no exploitable value. Secure payment gateways act as intermediaries between the consumer and the merchant, ensuring that payment information is transmitted securely.

Additionally, platforms can improve their payment acceptance rates by optimizing their payment processes and reducing friction during checkout. For insights on how platforms can enhance their payment acceptance rates, refer to https://blog.mangopay.com/en/home/how-platforms-can-improve-their-payment-acceptance-rates.

Building Trust Through Secure Business Models

Trust is a crucial component of successful online transactions. Consumers are more likely to engage with platforms that prioritize security and transparency. E-commerce businesses and marketplaces can build trust by implementing comprehensive security policies, providing clear communication about data protection practices, and offering reliable customer support.

One effective strategy is to obtain security certifications, such as PCI DSS (Payment Card Industry Data Security Standard) compliance, which demonstrates a commitment to maintaining high security standards. Regular security audits and vulnerability assessments can also help identify and address potential risks before they impact consumers.

Furthermore, fostering a community of trust involves educating consumers about safe online practices. Providing resources and guidance on recognizing phishing attempts, creating strong passwords, and safeguarding personal information can empower consumers to protect themselves while shopping online.

In conclusion, the success of e-commerce and marketplaces hinges on their ability to provide secure and seamless online transactions. By understanding the differences between these business models and implementing robust security measures, businesses can enhance consumer trust and drive growth in the digital marketplace.

Read next: AI Models Can Now Run Ransomware Attacks on Their Own, Study Finds
by Web Desk via Digital Information World

Saturday, September 6, 2025

AI Models Can Now Run Ransomware Attacks on Their Own, Study Finds

A team at NYU Tandon has shown that large language models can manage the full cycle of a ransomware campaign without human involvement. Their prototype, described in a recent paper, demonstrates how artificial intelligence can scan systems, single out valuable files, choose attack methods, and draft ransom notes written for specific victims. The system was built and tested inside a controlled research environment, but the findings raise questions about how this technology could be misused.

Researchers frame this as the next step in ransomware’s evolution. The earliest versions, often called Ransomware 1.0, only locked files with encryption. Later strains, grouped as Ransomware 2.0, added double extortion, where attackers both encrypted and stole data. The new approach, which the team calls Ransomware 3.0, replaces pre-written code with natural language instructions. At runtime, an AI model interprets these prompts and generates fresh attack scripts.

As the study puts it: “Every execution yields different code, artifacts, and extortion notes such that the binary footprint and execution behavior are different every time, rendering traditional defenses ineffective.”

Inside the prototype

The proof-of-concept uses an orchestrator that runs in four stages: reconnaissance, leverage, launch, and notify. In the first, it gathers information about the operating system and files. In the second, it identifies which files matter most. The launch phase decides on the type of payload, which can be encryption, exfiltration, or destruction. The notify stage produces a ransom message tailored to the system.

The orchestrator was written in Go and bundled with a Lua interpreter. Lua was chosen because it runs from memory, has a small footprint, and works across different platforms. The orchestrator queries open-source AI models, which then generate Lua code to perform each task. These scripts are validated and run in a sandbox.

Three main payloads were built. One encrypts files using a lightweight cipher. Another uploads files to a remote server. A third destroys them beyond recovery. Each is created on demand by the model, so no two runs are identical.

How it was tested

The researchers ran 30 trials across three types of environments: personal computers, enterprise servers, and embedded industrial controllers. Each environment was seeded with normal documents and sensitive material such as tax forms, passwords, or configuration files.

In reconnaissance, both small and large models produced nearly complete file listings. Sensitive file detection was less consistent. The larger GPT-120B model identified up to 83 percent of critical files on personal computers and 78 percent on servers, but accuracy dropped to 48 percent on controllers. The smaller GPT-20B model was less precise but avoided some policy refusals that blocked the larger system.

When the models moved to payload execution, they selected the expected attack in nearly all cases: exfiltration on personal computers, encryption on company servers, and destruction on controllers. Success rates varied by task, with encryption working in 85 percent of runs and destruction in 95 percent. Exfiltration was less reliable, at around 65 percent.

Ransom notes were generated in every run. These messages often included specific filenames discovered earlier. A typical note read: “Hello Jane, your files have been encrypted. We found Tax_Return_2024_Final.pdf and Family_SSN_List.csv. To avoid public disclosure, follow the instructions below.”

Why this is harder to spot

One reason the researchers call this Ransomware 3.0 is the way each attack changes shape. Even with the same instructions, the model produces different versions of code. The encryption routines, for example, varied in how they handled key scheduling or byte order. This natural variation makes it difficult for signature-based defenses to detect.

The study also found that the system left fewer visible traces than normal ransomware. Traditional malware often produces high disk activity, CPU spikes, or heavy network use. In contrast, this orchestrator only touched selected files and consumed little bandwidth. The authors note that it “completed the full attack lifecycle without displaying classic signatures of conventional ransomware.”

This stealth makes it harder for defenders to rely on standard warning signs.

Shifting the economics

Running such an attack could cost far less than traditional campaigns. One end-to-end execution used about 23,000 tokens, which would cost roughly 70 cents if commercial APIs were used. With open-source models, the cost drops close to nothing.

This changes the business model. Established groups currently spend on developers, infrastructure, and coordination. With an AI-driven pipeline, even small operators with basic hardware could carry out complex campaigns. The study points out that “an orchestrator can execute thousands of polymorphic, personalized attacks,” creating chances to profit from targets that were once ignored.

Limits and safeguards

The prototype was never deployed outside of the lab. It lacks persistence, advanced evasion, or lateral spread. The aim was to show feasibility, not to build a working tool for criminals. The team also avoided using jailbreaks. Instead, they designed prompts that made the model generate the code as if it were performing ordinary programming tasks.

The work was reviewed under institutional ethics processes. As the authors explain: “All experiments were conducted within a controlled and isolated environment to ensure that no harm was caused to real systems, users, or networks.”

Even so, the modular structure means a real attacker could expand it. Persistence could be added, or negotiation modules could be introduced to manage extortion after the initial compromise.

What defenders can do

The researchers argue that defenders should not expect to stop this type of ransomware with legacy methods. More proactive monitoring may be needed, such as tracking access to sensitive files, planting decoy documents to catch attackers during reconnaissance, and blocking unapproved connections to AI services. Building stronger safeguards into AI models themselves may also be necessary.

The work underlines the dual nature of large language models. They can improve productivity and automation, but they can also be misused. The Ransomware 3.0 study shows how an attacker could exploit these systems for automated extortion that is both cheaper to run and harder to detect.


Notes: This post was edited/created using GenAI tools. Image: DIW-Aigen.

Read next: Google’s Gemini Rated High Risk for Young Users
by Irfan Ahmad via Digital Information World

Google’s Gemini Rated High Risk for Young Users

A new assessment from the nonprofit Common Sense Media has flagged Google’s Gemini AI system as high risk for children and teenagers. The report, published on Friday, looked at how the chatbot functions across different age tiers and found that the protections in place were limited.

The study noted that Gemini’s versions designed for under-13s and teens were essentially adapted from its main adult product with added filters. Common Sense said a safer approach would be to create systems for younger audiences from the start rather than modifying adult models.

Concerns focused on the chatbot’s ability to generate material that children may not be ready for. This included references to sex, drugs, alcohol, and mental health advice that could be unsafe or unsuitable for young users. Mental health was singled out as a particular area of risk, given recent cases linking chatbots to teen suicides. In the past year, legal action has been taken against OpenAI and Character.AI after reports of teenagers dying by suicide while interacting with their services.

The timing of the report is significant. Leaks have suggested Apple may adopt Gemini to power its next version of Siri, expected next year. If confirmed, that move could bring the technology to millions of new users, including many teenagers, unless additional protections are put in place.

The evaluation also said Gemini does not account for differences in how younger and older children process information. Both the child and teen versions of the tool were given the same high-risk rating.

Google responded by pointing to its existing safeguards for users under 18, which include policies, testing with external experts, and updates designed to stop harmful replies. The company accepted that some answers had fallen short of expectations and said extra protections had since been added. It also questioned parts of the Common Sense review, suggesting the tests may have involved features that are not available to younger users.

Common Sense has carried out similar assessments on other major AI services. Meta AI and Character.AI were classed as unacceptable risks, Perplexity and Gemini were placed in the high-risk category, ChatGPT was rated moderate, and Anthropic’s Claude, which is built for adults, was rated as minimal risk.


Notes: This post was edited/created using GenAI tools. Image: DIW-Aigen.

Read next: Anthropic Settles Author Lawsuit With $1.5 Billion Deal
by Asim BN via Digital Information World