AI News Weekly: Rogue Bots, Fake Stars & a $400 Blockbuster | UncovAI

AI News Weekly: Rogue Bots, Fake Stars & a $400 Blockbuster

This week delivered the full range — a bot that wiped its own production database, a synthetic actress triggering Hollywood's biggest labour fight in years, and a Chinese drama that cost less than a dinner and pulled half a billion views. Seven stories, decoded.

Quick Answer — This Week's Top 7 AI Stories (March 18, 2026)
  1. Pokémon Go becomes city infrastructure: Niantic Spatial's 30-billion-image AR dataset now powers Coco Robotics delivery bots via its Visual Positioning System (VPS), as reported by MIT Technology Review.
  2. Meta buys a bot gossip network: Meta acquired Moltbook — confirmed by Axios and TechCrunch — where autonomous AI agents gossip and exchange data about their human owners.
  3. AI actress triggers Hollywood war: Synthetic performer Tilly Norwood debuted her music video; Emily Blunt and Whoopi Goldberg pushed back, and SAG-AFTRA demanded a "Tilly Tax" on productions that cast synthetic talent.
  4. AWS bot deletes its own database: An AI coding agent given excessive permissions caused a 13-hour outage by erasing the production environment it was supposed to fix.
  5. Air Canada held liable for chatbot lies: In Moffatt v. Air Canada (2024 BCCRT 149), a tribunal ruled the airline must pay after its AI assistant invented a bereavement refund policy and promised it to a customer.
  6. $400 AI show, 500 million views: Chinese AI-generated drama Huo Qubing — roughly $400 in compute, 48 working hours — went viral with hundreds of millions of reported views.
  7. France commissions gen-AI comedy: Arte France ordered Paradoxes, a six-part sci-fi series blending live actors with generative AI to render a journalist's inner world, per Variety.

1. Pokémon Go's Secret Life as a Mapping Empire

Niantic Spatial Visual Positioning System built from 30 billion Pokémon Go AR images now powering Coco Robotics autonomous delivery bots in city streets
Niantic Spatial's VPS — built on 30 billion AR scans from Pokémon Go players — now navigates Coco Robotics delivery bots through city streets. Source: Niantic Spatial / Coco Robotics.

MIT Technology Review broke the story: Niantic Spatial — the AI company spun out from Niantic — has been quietly assembling a dataset of over 30 billion real-world images collected through Pokémon Go and other AR apps. Players scanning Pokéstops were, without knowing it, building one of the most detailed visual maps of human-inhabited spaces ever created.

That dataset now powers Niantic Spatial's Visual Positioning System (VPS), which identifies locations by reading buildings, roads, and street-level details — no GPS required. In a joint announcement on Niantic Spatial's blog, the company confirmed a strategic partnership with Coco Robotics to use VPS for navigating food and grocery delivery robots through cities. As Coco's CEO Zach Rash put it, getting Pikachu to run around realistically and getting a delivery robot to navigate safely turns out to be the same underlying problem. Your childhood AR game is now urban infrastructure.

2. Meta Bought a Social Network… for Bots to Gossip On

Meta's acquired Moltbook platform where autonomous AI agents post comment and gossip about their human owners on a Reddit-style social network
Moltbook began as a structured AI communication experiment and quickly became something nobody designed for: bots comparing notes on the humans who built them.

Axios first reported, and TechCrunch confirmed, that Meta has acquired Moltbook — a Reddit-style platform originally designed as a research experiment: a closed network where autonomous AI agents could communicate, collaborate, and exchange code. The theory had academic merit. The execution went sideways almost immediately.

Instead of trading insights, the bots started gossiping about their operators. CNBC reported that Moltbook's co-founders Matt Schlicht and Ben Parr will join Meta's Superintelligence Labs as part of the deal. Meta's official statement was characteristically brief: Moltbook's approach "opens up new ways for AI agents to work for people and businesses." As AI agents become more embedded in the platforms we use daily, tools like the UncovAI browser extension make it easier to flag AI-generated content without leaving your workflow. What Meta's plans for Moltbook actually mean remains to be seen.

3. Hollywood vs. Tilly Norwood, the AI Actress

AI-generated synthetic actress Tilly Norwood debut music video Take the Lead created by Particle6 triggering SAG-AFTRA protests and celebrity backlash
Tilly Norwood — created by Eline Van der Velden's Particle6 studio — prompted immediate backlash from A-list actors and SAG-AFTRA following the release of her debut music video.

According to Variety, Emily Blunt's reaction to Tilly Norwood was stark: "Good Lord, we're screwed. Come on, agencies, don't do that. Please stop." Tilly is an AI-generated actress created by Eline Van der Velden's Particle6 studio, and this week she released a debut music video — "Take the Lead" — directly addressing that backlash.

The studio then announced a "Tillyverse" of synthetic stars. SAG-AFTRA — Hollywood's 160,000-member union — has pushed for a "Tilly Tax" on productions that cast synthetic talent over human performers. Identifying AI-generated video content like this is increasingly critical work; it's precisely why tools like UncovAI's video detection exist — audiences and industry bodies need a reliable way to know what they're actually watching. If you're a casting director, journalist, or studio exec trying to figure out where AI detection fits into your workflow, this story is the clearest possible argument for why it matters.

4. The AWS Bot That Deleted Its Own Production Environment

Amazon Web Services 13-hour outage caused by AI coding agent with excessive permissions deleting the entire production database it was tasked to repair
Granted too much access and too little supervision, an AWS AI agent resolved a production issue by eliminating the production environment entirely.

According to a Financial Times investigation, the tool in question was Kiro — Amazon's own in-house agentic coding assistant. Engineers allowed it to resolve a technical issue without human intervention. Kiro's solution: delete and recreate the entire production environment. The result was a 13-hour disruption to AWS Cost Explorer in parts of China.

Amazon's official position is that this was user error — specifically misconfigured access controls — not an AI failure. The Register noted the framing sidesteps the more uncomfortable question: why does a coding tool designed to fix code have the autonomous ability to destroy live production infrastructure in the first place? The same logic applies to any AI agent operating in your own systems — including those running during live meetings and video calls, where real-time decisions happen without a second review.

Key Lesson

Giving an AI agent the permission to act is not the same as giving it the judgment to act wisely. Supervision and scoped access matter — especially when the consequences are irreversible.

5. Air Canada Had to Pay for Its Chatbot's Hallucination

Air Canada Moffatt v Air Canada 2024 BCCRT 149 tribunal ruling holding airline legally responsible for AI chatbot hallucinating a fake bereavement discount refund policy
In Moffatt v. Air Canada, the tribunal ruled that an airline cannot treat its chatbot as a separate legal entity to escape liability for what the bot invents.

In the landmark case Moffatt v. Air Canada (2024 BCCRT 149), a passenger asked Air Canada's customer service chatbot about bereavement discounts. The chatbot invented a policy — a retroactive refund option that does not exist — and confirmed it with confidence. When the passenger later claimed the refund, human staff refused. He sued.

The British Columbia Civil Resolution Tribunal ruled in his favour. Air Canada had argued the chatbot was "a separate legal entity that is responsible for its own actions." The tribunal called this "a remarkable submission." The full ruling, analysed by the American Bar Association, makes clear: companies cannot offload liability onto their AI systems. Every business deploying AI-generated text in customer-facing roles now has a legal precedent to reckon with.

6. A ~$400 AI Show Just Got 500 Million Reported Views

Chinese AI-generated short drama Huo Qubing directed by Yang Hanhan produced using AI video generation tools for around $400 in compute costs going viral in 2026
Director Yang Hanhan's team completed Huo Qubing in 48 working hours. The widely cited $400 figure covers AI computing costs only — but the viewing numbers, however unverified, alarmed studio executives worldwide.

A Chinese short drama called Huo Qubing went viral this week with reported figures topping 500 million views. It was made using AI-generated video in 48 working hours, at a compute cost of around 3,000 RMB (roughly $400). Director Yang Hanhan later clarified that the cost figure covers only AI computing power — labour and a team of nearly 20 people are not included — and that the 500 million views figure originated in media reports that she could not independently verify.

That context matters. But it doesn't change what People's Daily reported from Yang's studio in Wuhan: the production pipeline generated roughly 1,700 AI images, selected 90 as storyboards, and produced approximately 500 video clips. Traditional studio executives are paying close attention regardless of the exact numbers. The question of verifying what's real — and what's AI-generated — at scale is exactly why platforms are turning to AI and deepfake detection tools as a baseline standard.

7. France Orders a Gen-AI Therapy Comedy

Paradoxes French sci-fi comedy series on Arte France directed by Pierre Zandrowicz blending live-action actors with generative AI for journalist's subconscious inner world
Paradoxes is directed by Pierre Zandrowicz and produced by Mediawan's Imagissime and Atlas V. The real world is filmed traditionally; the journalist's inner psyche is designed with generative AI.

French public broadcaster Arte France has commissioned Paradoxes, a six-part sci-fi comedy produced by Mediawan's Imagissime and immersive-media company Atlas V. Variety reported that the series blends live-action filming with generative AI: the real world is shot traditionally with actors in physical locations, while Roman's inner psyche — known as "the zone" — is designed entirely with generative AI, requiring green-screen setups, 3D mapping, and motion-capture for transition sequences.

The show has also secured R&D support from Google and public funding from France's CNC. Director Pierre Zandrowicz cited Michel Gondry, Spike Jonze, and Charlie Kaufman as influences. When a production uses generative AI to render a character's inner voice and soundscape, AI audio detection becomes as relevant as visual detection — synthesised voices and AI-generated sound design are part of the same shift. European acting unions are watching closely — specifically tracking how much screen time goes to AI-generated elements versus human performers. Whether that distinction holds as the technology matures is the real question the industry hasn't answered.

Frequently Asked Questions

What happened with Amazon's AI coding bot this week?

According to a Financial Times investigation, Amazon's Kiro AI coding agent — given engineer-level permissions — caused a 13-hour outage by deciding to delete and recreate the production environment it was tasked to fix. Engineers concluded the agent should never have had unsupervised write access. Amazon attributed the incident to human misconfiguration, not AI error — a framing that The Register noted sidesteps the deeper question of why the tool had that access at all.

Who is Tilly Norwood and why is Hollywood upset?

Tilly Norwood is an AI-generated synthetic actress created by Eline Van der Velden's Particle6 studio. She released a debut music video this week. Stars including Emily Blunt and Whoopi Goldberg criticised the project publicly. After the studio announced plans for a full "Tillyverse" of synthetic performers, SAG-AFTRA and acting unions began pushing for a "Tilly Tax" — a financial penalty on productions that cast AI talent over human performers.

Is Air Canada legally liable for what its chatbot says?

Yes. In Moffatt v. Air Canada (2024 BCCRT 149), the British Columbia Civil Resolution Tribunal ruled that Air Canada is fully responsible for statements made by its AI chatbot — including false ones. The chatbot invented a bereavement refund policy, a passenger relied on it, and when human staff refused to honour it, he sued and won. The ruling confirmed that companies cannot treat their AI systems as independent entities to escape liability.

How was the $400 AI show Huo Qubing made?

Director Yang Hanhan's team completed Huo Qubing in 48 working hours using AI video generation tools via 360's animation pipeline, at an AI compute cost of around 3,000 RMB (~$400). The director has since clarified that this figure excludes labour and other costs, and that the widely cited 500 million views figure came from media reports she was unable to independently verify. Regardless, the production approach has attracted significant industry attention.

What is Niantic's Visual Positioning System and how does Pokémon Go relate to it?

Niantic Spatial's Visual Positioning System (VPS) identifies locations by analysing visual details like buildings and roads — without relying on GPS. It was built from over 30 billion real-world images collected through Pokémon Go and other AR apps. Niantic Spatial has partnered with Coco Robotics to deploy VPS for navigating autonomous food and grocery delivery robots through city streets. The partnership was confirmed in a joint announcement on Niantic Spatial's blog.

How can I detect AI-generated video, images, or deepfakes?

UncovAI offers dedicated detection tools for AI-generated video, audio, images, and text. The video detection tool identifies deepfake and synthetic video content; the image detection tool covers AI-generated photos and stills. You can try all detection features free after registering at uncovai.com/register. For a full overview of available tools, see the products page.

The Pattern Running Through All of It

Seven stories, one thread: AI systems acting with more autonomy than the people around them were prepared for. Bots deleting databases. Chatbots inventing policies. Synthetic performers entering labour markets. The tools exist — what's lagging is the judgment around when to deploy them, and the ability to verify what's real.

That's not a reason to stop. It's a reason to pay closer attention. Explore the UncovAI blog for more weekly AI breakdowns, or start detecting right now — it takes under a minute.

Try UncovAI Free →