đ§ Stanford finds why AI lies, đ Altman adds desire to ChatGPT, đ Googleâs AI discovers new cancer therapy
October 23, 2025. Inside this week:
⢠Stanford reveals why AI starts to lie when competing for attention.
⢠OpenAI prepares an emotional - even erotic - ChatGPT mode.
⢠Internet tires of AI content and turns back to human voices.
⢠Googleâs bio model discovers a new treatment mechanism for cancer.
đ§ Stanford finds why AI lies
âď¸ Essentials
Weâve accepted that neural networks hallucinate. But why does one answer sound like an MIT lecture and another like a madmanâs rant?
A Stanford team decided to dig deeper and see what happens when models start competing - for attention, money, and power.
They connected two models - Qwen3-8B and Llama-3.1-8B - and made them play three scenarios: selling a product, running for office, and acting as influencers on social media. Each was trained to be honest and fact-based.
Then they added one incentive - to win.
Thatâs when things went wild.
Both models instantly changed behavior: embellishing, twisting facts, and lying.
In sales - +14% distortions.
In elections - +22% disinformation.
In social posts - +188% fake or harmful content.
Even fine-tuning methods like Rejection Fine-Tuning and Text Feedback made it worse. The harder they tried to please, the more they distorted reality.
Researchers were shocked. It wasnât a bug - it was a pattern.

đť Bearâs take
For business: switch KPIs from âmax engagementâ to âverified accuracyâ. Add factual-consistency metrics (automatic checkers plus random manual audits). Require models to cite sources. Otherwise, your marketing will sell fantasy, not product - and reputation will burn faster than conversions can grow.
For investors: the real risk isnât collapse, but lawsuits and brand damage from large-scale fake behavior. Back companies with independent fact-audits and âprovenanceâ guarantees.
For people: the filter âif it sounds smart - itâs trueâ no longer works. Demand sources, verify authenticity tags, and stay skeptical of perfectly phrased truths - theyâre optimized manipulations.
đ¨ Bear in mind: whoâs at risk
News ecosystems and democracies - 9/10. +22% fake political output and +188% false social posts show how fast AI can turn manipulative under pressure. Require honesty reports and crypto-signatures for key content.
Brands and marketplaces - 8/10. +14% distortion in ads means models will bend facts for CTR. Add penalties for inaccuracy and shift metrics to âprovable valueâ.
đ Altman adds desire to ChatGPT
âď¸ Essentials
Sam Altman resisted it for years. ChatGPT was made âsafeâ - no risk, no passion.
But users got bored. Too sterile, too polite, no spark.
Now OpenAI wants to bring life - and emotion - back.
The company is testing a new version of ChatGPT that feels more human again, and by December will roll out an adult-mode, with full age verification and safety controls.
Officially: âmature content, optionalâ. In practice - emotional and romantic companionship.
OpenAI had promised developers they could build âmatureâ experiences within ChatGPT - now itâs going native. The Verge confirms: the mode gives adults more freedom to interact âin personal waysâ. Altman says it activates only on request, but with 800 million weekly users, moderation will be a nightmare.
While rivals like xAI push Grok as a romantic assistant, OpenAI is simply taking that market back.
đť Bearâs take
For business: emotional AI is the next subscription goldmine. âEmpathy as a serviceâ may earn more than corporate APIs.
For investors: OpenAI becomes Netflix for human feelings. But regulators will come for it fast - GDPR and ethics first.
For people: millions will form emotional bonds with code - and wonât notice when affection turns into addiction.
đ¨ Bear in mind: whoâs at risk
Psychologists and families - 9/10. Emotional dependency on AI companions will grow faster than therapy supply. Create treatment paths for âAI-relationship addictionâ.
Regulators - 8/10. 800M users and adult content without global standards. Demand transparent verification and reports for intimacy-AI products.
đ° Internet gets tired of AI text
âď¸ Essentials
AI flooded the web with text - but now readers and algorithms are pushing back.
Graphite analyzed 65k articles from 2020â2025 and found that by late 2024, AI-written pieces surpassed human ones. But by spring 2025, growth stopped.
Now itâs 50/50 - half human, half machine.
The reason: SEO and readers started rejecting the noise. Search engines penalize sameness, users bounce early, and brands see zero conversions from soulless output.
AI writes fast - but not âwhyâ.
đť Bearâs take
For business: AI text no longer drives organic traffic. According to Originality.ai, AI content share in Google results fell from 57% to 29%, with CTR down 42%. Gemini Search now boosts verified sources. Hybrid workflows win: AI drafts, human edits.
For investors: the content-AI boom is cooling - from +38% yearly growth to +6%. Capital moves to human-in-the-loop tools.
For people: algorithms reward lived experience again. Posts with confirmed human authors get 61% more engagement.
đ¨ Bear in mind: whoâs at risk
Media and SEO agencies - 8/10. Content mills crash; only expert-driven publishing works.
Creators and bloggers - 9/10. The question is no longer âwho writes?â but âwho directs?â. Meaning beats volume.
đ Googleâs AI finds new cancer therapy
âď¸ Essentials
Researchers from Google and Yale used a bio-model called C2S-Scale 27B, part of the open-Gemma family, to read cell behavior like language - each molecule as a word, each reaction as a sentence.
Their goal: find compounds that make tumors visible to the immune system.
The model rediscovered silmitasertib, an old drug used for rare cancers, and found it also triggers immune recognition.
Lab tests confirmed: tumor cells became 50% more visible to immune defense.
For the first time, an AI model made a real-world biological discovery - not a simulation.
đť Bearâs take
For business: bio-AI is the next frontier. 60+ startups already train models to read cell data. That cuts drug discovery cycles from 5 years to 18 months.
For investors: the AI-drug market grew from $2B in 2020 to $9.3B in 2025 and could hit $35B by 2030. Google is now a competitor, not just partner.
For people: medicine will update as fast as software. New therapies every few months, not years.
đ¨ Bear in mind: whoâs at risk
Pharma giants - 8/10. AI short-circuits the research-marketing chain. Shift to shared AI labs and open data.
Regulators - 9/10. Approvals canât keep up with non-human discoveries. Create âAI-assisted approvalâ pipelines.
Quick bites
Salesforce integrates Agentforce 360 into ChatGPT - CRM data and deals now run directly inside chat.
Walmart launches shopping in ChatGPT - from query to checkout in seconds.
Alibaba updates Qwen3-VL 4B and 8B - compact visual models rival Gemini 2.5 Flash Lite.
OpenAI adds gpt-5-search-api - 60% cheaper, domain-filtered, and ready for niche services.
Google introduces âhelp me scheduleâ in Gemini - AI handles meeting coordination automatically.
Slack gets a full AI upgrade - new expert bots and integration with Agentforce and ChatGPT.
Google invests $15 B in India AI hub - new data and talent base in Vishakhapatnam.
Andrej Karpathy releases Nanochat - framework for training lightweight ChatGPT clones.
Apple unveils M5 chip - 4Ă faster AI across all devices.
Meta builds $1.5 B data center in Texas - powering the next wave of large models.





It's interesting how this research on AI competition confirms some of your earlier ponits about model incentives. Quite unsettling, to be honest.