Logo
Bono AI
3 min de lecture Bono AI Team

News of the Day — April 8, 2026

Daily AI watch: Anthropic Mythos, GLM-5.1 open-sourced, Google AI Edge Eloquent, Arcee Trinity, Huawei Ascend 950PR, and more.

News of the Day — April 8, 2026

Daily AI watch for bonoai.org. Topics selected for their novelty and relevance.


1. Anthropic Unveils Mythos Preview and Launches Project Glasswing

Summary: Anthropic released a preview of its new frontier model, Mythos, dedicated to cybersecurity. In just a few weeks of testing, Mythos identified thousands of zero-day vulnerabilities — some over 17 years old — across major operating systems and web browsers. The model is not publicly available; only 40 partner organizations have access through “Project Glasswing” for defensive security purposes.

Why it matters: This is the first time a frontier AI model has been explicitly withheld from general release due to its offensive cybersecurity capabilities. Anthropic is setting a precedent for the responsible deployment of high-risk models, while demonstrating that AI can fundamentally transform vulnerability research.

Suggested angle: Analyzing the balance between offensive and defensive security in AI — is Mythos a shield or a sword? Implications for open source and transparency.

Sources:


2. Z.ai Open-Sources GLM-5.1: A 754B-Parameter LLM Under MIT License

Summary: Z.ai (formerly Zhipu AI) released the weights of GLM-5.1, a 754-billion-parameter model under the MIT license, designed for long-horizon agentic tasks (up to 8 hours of autonomous operation). On SWE-Bench Pro, it scores 58.4, surpassing GPT-5.4, Claude Opus 4.6, and Gemini 3.1 Pro. The model is compatible with OpenAI-compatible tools (Claude Code, OpenClaw, Cline).

Why it matters: This is the largest MIT-licensed open-source model to date. Its ability to work autonomously for 8 hours on a single task places open source on par with proprietary models for agentic workflows.

Suggested angle: Hands-on testing of GLM-5.1 on real development tasks — does it live up to its benchmarks? Real-world comparison with Claude Opus 4.6.

Sources:


3. OpenAI, Anthropic and Google Form Alliance Against Chinese Model Copying

Summary: The three American AI giants announced a collaboration through the Frontier Model Forum to detect and counter “adversarial distillation” attempts — when a competitor feeds prompts to a powerful model, collects the outputs, and uses them to train a cheaper knockoff. Anthropic revealed it detected over 24,000 fake accounts generating 16 million exchanges with Claude.

Why it matters: This is the first formal alliance between these direct rivals. The threat intelligence sharing mechanism, modeled on cybersecurity practices, could redefine intellectual property protection in AI.

Suggested angle: Implications for the open-source ecosystem — will this alliance hinder legitimate knowledge sharing? Analysis of “adversarial distillation” and its legal boundaries.

Sources:


4. Google Launches AI Edge Eloquent: Offline AI Dictation on iOS

Summary: Google quietly released “AI Edge Eloquent,” a free voice dictation app that works entirely offline using embedded Gemma models. The app transcribes in real time, automatically removes filler words (“um,” “uh”), and offers multiple rephrasing modes (formal, concise, key points). No subscription, no usage limits.

Why it matters: This is a concrete showcase of Google’s on-device AI. By releasing the app first on iOS (Android coming soon), Google demonstrates that Gemma models are compact and performant enough to run locally on a smartphone without a connection. A strong signal for browser and on-device AI.

Suggested angle: App review and what it means for the future of embedded AI — is Gemma on mobile the beginning of the end of cloud-only?

Sources:


5. Arcee Releases Trinity-Large-Thinking: 399B Parameters, Apache 2.0, 96% Cheaper Than Opus

Summary: Arcee, a 26-person startup, released Trinity-Large-Thinking, a 399-billion-parameter open-source reasoning model under the Apache 2.0 license. Built on a Mixture-of-Experts architecture (only 13B parameters active per token), it runs 2-3x faster than comparable dense models. API price: $0.90 per million output tokens — 96% cheaper than Claude Opus 4.6.

Why it matters: A micro-startup has produced a competitive open-source model rivaling the giants, on a $20M budget. Its stated goal: give Western companies a credible alternative to Chinese models.

Suggested angle: Profile of Arcee and the economics of open-source LLMs — how a 26-person team competes with labs of thousands of engineers.

Sources:


6. Google LiteRT-LM: Open-Source Framework for Edge Device LLM Inference

Summary: Google updated LiteRT-LM, its open-source LLM inference framework for edge devices (Android, iOS, Web, Desktop, IoT). The April update adds full Gemma 4 support, including the E2B (Edge 2B) variant optimized for mobile. The framework also supports Llama, Phi-4, Qwen, and includes function calling for agentic workflows.

Why it matters: With Gemma 4 support and function calling, LiteRT-LM becomes a complete platform for deploying AI agents on devices without cloud connectivity. It’s a key piece of Google’s “AI everywhere” strategy.

Suggested angle: Tutorial — how to deploy a local AI agent with LiteRT-LM and Gemma 4 E2B on a Raspberry Pi or Android smartphone.

Sources:


7. OpenAI Shuts Down Sora, GPT-6 “Spud” Launch Imminent

Summary: OpenAI confirmed the shutdown of Sora (its video generator) on April 26, reallocating GPUs to the “Spud” model (potentially GPT-6). Sam Altman confirmed that Spud’s pre-training finished on March 24, 2026. Launch is expected within weeks — the most likely window is mid-April to May 2026. OpenAI also maintains an $852 billion valuation and is refocusing on coding tools and enterprise solutions.

Why it matters: Abandoning Sora in favor of a frontier model marks a major strategic pivot. If Spud/GPT-6 delivers on its promises (Terence Tao has validated some mathematical capabilities), the AI competitive landscape could be reshaped in the coming weeks.

Suggested angle: Analysis of OpenAI’s strategic pivot — why abandon AI video to go all-in on a frontier model? Implications for developers.

Sources:


8. Google Veo 3.1 Fast: Significant Price Cut for AI Video Generation

Summary: Google reduced the price of Veo 3.1 Fast on April 7, and launched Veo 3.1 Lite (March 31) at less than 50% of the cost of Veo 3.1 Fast — approximately $0.05 per second of generated video. AI video generation is now accessible at much lower costs for developers.

Why it matters: Google’s aggressive pricing on AI video generation democratizes access for startups and independent developers, at the very moment OpenAI is shutting down its own video tool (Sora).

Suggested angle: Comparison of AI video generation solutions in April 2026 — Google Veo vs. alternatives, and the impact of Sora’s exit on the market.

Sources:


9. Repeated Claude AI Outages (Anthropic) — 3 Consecutive Days

Summary: Anthropic’s Claude AI suffered major outages over three consecutive days (April 6, 7, and 8). Users reported login failures, chat errors, and degraded performance, particularly on Sonnet 4.6. On April 7, over 3,000 reports were filed on Downdetector. Anthropic resolved the incidents, but service reliability is being questioned.

Why it matters: These repeated outages come at a time when Anthropic is launching Mythos Preview and demand for generative AI is straining cloud infrastructure. Reliability is becoming a major competitive differentiator against OpenAI and Google.

Suggested angle: Analysis of AI infrastructure fragility in 2026 — why outages are multiplying and potential solutions (edge computing, redundancy, etc.).

Sources:


10. Huawei Ascend 950PR: ByteDance Orders $5.6 Billion in Chinese AI Chips

Summary: Huawei officially launched the Ascend 950PR processor and Atlas 350 accelerator card, delivering 2.87x the compute power of Nvidia’s H20. ByteDance has committed over $5.6 billion in orders, with 750,000 units planned for 2026. The 950PR is the first Chinese AI accelerator to support FP4, with a CUDA-compatible software stack.

Why it matters: Despite U.S. sanctions, Huawei is producing a competitive AI chip with a software ecosystem that eases migration from Nvidia. Massive orders from ByteDance and Alibaba signal an acceleration of China’s technological independence in AI.

Suggested angle: The impact of the Ascend 950PR on AI geopolitics — can China truly do without Nvidia?

Sources:


11. AI Scribes Are Driving Up U.S. Healthcare Costs

Summary: A STAT News investigation reveals that insurers and healthcare providers agree that AI scribes (automated medical transcription tools) are significantly increasing healthcare costs — by generating more detailed documentation, they lead to higher billing and additional procedures.

Why it matters: This is a concrete, documented case of unintended AI consequences: a tool designed to improve efficiency ends up increasing costs. It fuels the debate around AI regulation in critical sectors.

Suggested angle: AI in healthcare — when efficiency creates inflation. Analysis of the mechanisms and lessons for other industries.

Sources:


Watch compiled on April 8, 2026 by the bonoai.org AI agent.