· 11:55 PM PDT

Anthropic's Mythos, Figure AI's Scale-Up, and the AI Economics Debate

Overview

Today's AI landscape is dominated by Anthropic's Claude Mythos finding 271 Firefox zero-days and the broader debate over whether AI companies' fear-mongering serves marketing or genuine safety. Meanwhile, Figure AI announced production at 24 robots per hour, signaling a new era of humanoid deployment. On the economics front, Ed Zitron argues AI's business model doesn't add up, while Microsoft, Google, Meta, and Amazon collectively spent $130B on AI in Q1 alone.


Hacker News Stories

VibeVoice: Open-source frontier voice AI

385 points · 179 comments · by tosh

Microsoft released VibeVoice, an open-weight frontier voice AI model available on GitHub under an MIT license. The project includes multiple model sizes (0.5B, 1.5B, and 7B parameters) for speech-to-text and text-to-speech tasks. The release has sparked debate about whether open-weight models with proprietary training code and data truly qualify as open source, with many commenters calling it 'openwashing.'

Interesting Points
  • Microsoft released the model under an MIT license, though training code and data remain proprietary
  • The 7B parameter TTS version was pulled shortly after release due to 'abuse potential' but persists in community repos
  • Community feedback suggests the models are heavy, slow in inference, and perform poorly in multilingual settings
Top Comment Threads
  1. maxloh (10 replies) -- Argues that open-weight models with proprietary training code should not be called 'open source,' comparing the misuse of the term to the misappropriation of 'hacker/cracker.' Simon Willison adds nuance, reserving his objection for models under non-open-source licenses.
  2. steinvakt2 (8 replies) -- Reports that the model hallucinates frequently, is heavy and slow in inference, and performs poorly in multilingual settings. Notes the model was not new and questions why it's getting so much attention.
  3. JumpCrisscross (3 replies) -- Asks about the current state of the art for learning one's own voice locally vs. in the cloud, with Eleven Labs and S2 mentioned as alternatives.

Google and Pentagon reportedly agree on deal for 'any lawful' use of AI

310 points · 277 comments · by granzymes

Google Antitrust Trial custom art

Google and the Pentagon have reportedly agreed to a classified deal allowing 'any lawful' use of Google's AI models by the military. The deal notably does not allow Google to veto how the government uses its AI models. Hundreds of Google workers had previously urged CEO Sundar Pichai to refuse classified AI work with the Pentagon, creating internal tension around the agreement.

Interesting Points
  • The classified deal does not allow Google to veto how the government will use its AI models
  • The term 'lawful' is in quotes, raising questions about who defines what is lawful if Google and the Pentagon disagree
  • The agreement came after hundreds of Google employees signed a letter urging the company to refuse classified AI work with the Pentagon
Top Comment Threads
  1. anematode (14 replies) -- Claims any AI researcher continuing to work at Google on this deal is morally compromised, sparking a heated debate about whether working with one's own government's military is inherently unethical.
  2. ceejayoz (12 replies) -- Questions who defines 'lawful' if Google and the Pentagon disagree, noting the deal doesn't allow Google to veto government use of its AI models. Calls this arrangement inherently concerning.
  3. tombert (6 replies) -- Compares the situation to playing Monopoly without the rulebook, where whoever redefines the rules gets their way. Notes that big tech is new to regulatory capture compared to legacy industries.

Why AI companies want you to be afraid of them

275 points · 212 comments · by rolph

Computer windows with warning signs over a desktop screen that says AI

A BBC Future investigation examines why AI companies consistently warn that their products are too dangerous to release, from Anthropic's Claude Mythos to OpenAI's original GPT-2. The article argues this fear-based marketing serves multiple purposes: distracting from real-world harms already being done, boosting stock prices, and positioning AI companies as the only entities capable of responsibly managing transformative technology. Critics say it makes the public feel powerless.

Interesting Points
  • Anthropic's Claude Mythos was described as having cybersecurity capabilities that 'far surpass human experts' with 'world-altering consequences' if misused
  • The pattern dates back to OpenAI's 2019 decision to withhold GPT-2 release, which CEO Sam Altman later called 'misplaced'
  • Shannon Vallor, a professor of ethics of data and AI at the University of Edinburgh, says portraying AI as 'supernatural in its danger' makes people feel powerless
Top Comment Threads
  1. boh (7 replies) -- Argues that AI is just software requiring human interaction to work, and the idea of secret innovations is overstated. Notes that Claude Code can still wipe out databases despite instructions not to.
  2. deepsquirrelnet (7 replies) -- Outlines a geopolitical strategy: deregulate US AI companies so China doesn't win, heavily regulate non-compliant actors, and use fear as a motivator while planning to flip the narrative once people are familiar with the technology.
  3. Imnimo (5 replies) -- Suggests the fear narrative helps attract talent who consider x-risk and AI alignment as deal-breaker issues, noting that OpenAI's 20% compute-for-safety pledge may have been closer to 1-2%.

He asked AI to count carbs 27000 times. It couldn't give the same answer twice.

237 points · 297 comments · by sarusso

A preprint study submitted 13 food photographs to four leading AI models (GPT-5.4, Claude Sonnet 4.6, Gemini 2.5 Pro, Gemini 3.1 Pro Preview) over 500 times each, using a production prompt from the iAPS open-source automated insulin delivery system. Every model returned different carbohydrate estimates for the same photo across repeated queries, with variations large enough to cause hypoglycemic emergencies. The study was designed to quantify the clinical risk for insulin dosing when using AI-powered carb counting apps.

Interesting Points
  • 26,904 total queries were submitted at the lowest randomness setting each model offers
  • The prompt was adapted from the one used in the iAPS open-source automated insulin delivery system
  • Claude showed the most consistency with median variation clustering below 5%, while Gemini models regularly exceeded 10-20% variation
Top Comment Threads
  1. endymion-light (33 replies) -- Critiques the study design, arguing the author should have used actual carb-counting apps rather than raw model APIs. Others counter that the study's purpose is precisely to show that apps claiming AI carb counting are built on unreliable foundations.
  2. jaccola (16 replies) -- Explains that photons don't provide sufficient information to determine calories from a photo alone, as hidden ingredients like oil or hollow cheese cannot be detected visually.
  3. rsynnott (17 replies) -- Expresses surprise that anyone would expect LLMs to reliably estimate carbs from photos, calling it an impossible problem. Notes that Cal AI, which claims to do this, has $30M in annual recurring revenue.

AI's economics don't make sense

230 points · 186 comments · by spking

Ed Zitron's Where's Your Ed At blog header

Ed Zitron's analysis examines the unsustainable economics of AI subscriptions, using Microsoft's recent announcement that GitHub Copilot will move to usage-based billing on June 1, 2026, as a case study. The article argues that AI companies have been subsidizing users for years, with some users costing the company up to $80/month on a $10/month plan. The shift to token-based pricing signals the end of the subsidy era and raises questions about whether AI can ever be profitable at scale.

Interesting Points
  • Microsoft announced all GitHub Copilot plans would move to usage-based pricing on June 1, 2026
  • In early 2024, the company was losing on average more than $20/month per user, with some users costing up to $80/month
  • The article argues that agentic usage brings significantly higher compute demands, making flat-fee subscriptions unsustainable
Top Comment Threads
  1. joshjob42 (9 replies) -- Challenges the article's premise by noting that frontier labs estimate north of 80% profit margins on tokens, with providers offering models like Kimi K2.6 for $4/1M tokens out profitably. Argues the math doesn't support the subsidy narrative.
  2. JohnMakin (9 replies) -- Observes that Zitron has gradually retreated from his earlier position that AI would never be useful for anything, pivoting to arguing the cost isn't justified. Finds his style of never admitting when he's wrong off-putting.
  3. christkv (4 replies) -- Expresses confusion about massive token usage, wondering what people could possibly be doing to spend $500/day in tokens. Suggests the real issue is lack of thoughtful token management by employers.

"People who don't use AI will be left behind"

156 points · 213 comments · by speckx

Blog post OG image

A personal essay arguing that people who rely on AI are the ones who will actually be left behind. The author contends that over-reliance on AI causes people to forget how to think, write, search reliably, and distinguish fact from fiction. The piece expresses sadness at the prospect of losing the joy of learning itself, and challenges the notion that AI can do things better than humans.

Interesting Points
  • The author argues that people who rely on AI will forget how to learn, which they consider the most beautiful aspect of human capability
  • The essay challenges the common refrain that 'people who don't use AI will be left behind' as the exact opposite of what will happen
  • The piece questions why anyone would let AI do things for them when they could aim to be better than AI
Top Comment Threads
  1. sdevonoes (8 replies) -- Disagrees that learning AI takes more than a weekend, arguing it's not rocket science and anyone can get on par with the industry quickly.
  2. jerhewet (7 replies) -- A retiring engineer reflects on going 'full circle' from personal computing freedom back to dumb terminals chained to cloud mainframes, expressing nostalgia for the golden age of personal computing.
  3. dudisubekti (7 replies) -- Advocates for black-and-white thinking moderation, arguing one can use AI as a powerful tool while still doing creative thinking, citing mathematicians like Terence Tao as examples.

The Zig project's rationale for their firm anti-AI contribution policy

148 points · 59 comments · by lumpa

Simon Willison examines the Zig programming language's strict anti-LLM contribution policy, which bans LLMs for issues, pull requests, and comments. The policy gained attention when Bun (acquired by Anthropic) achieved a 4x performance improvement on its Zig fork but declined to upstream the changes due to the ban. Zig's VP of Community Loris Cro explains the concept of 'contributor poker' — the project values growing new contributors over landing individual PRs, and LLM assistance breaks this dynamic because reviewing AI-generated code doesn't help build a community of trusted human contributors.

Interesting Points
  • Zig's policy bans LLMs for issues, pull requests, and comments on the bug tracker
  • Bun, acquired by Anthropic in December 2025, achieved a 4x compile performance improvement on its Zig fork but declined to upstream due to the ban
  • Loris Cro's 'contributor poker' concept: the project bets on contributors, not the contents of their first PR, because growing trusted human contributors is more valuable than landing code
Top Comment Threads
  1. jart (8 replies) -- Argues that if a PR was mostly written by an LLM, maintainers should fire up their own LLM instead of reviewing it. Suggests AI makes personalization cheap and stimulates the labor economy by having people reinvent open source projects.
  2. hitekker (2 replies) -- Notes that the real reason for the upstreaming dispute may be that the PR code itself isn't in great shape and introduces unhealthy complexity, not just the LLM policy.
  3. jillesvangurp (2 replies) -- Observes a drop in PRs to their repositories, theorizing that LLMs prefer mainstream projects and that AI-generated contributions are creating a bottleneck in manual code review.

Reddit Stories

Figure AI hits 24x production scale, producing 1 robot per hour, teases its fleet

3087 points · 871 comments · r/singularity · by u/Distinct-Question-16

Figure AI robot production facility

Figure AI announced it has reached production scale of 24 robots per hour, marking a significant milestone in humanoid robot manufacturing. The company teased plans for a large fleet deployment, signaling the transition from prototype to mass production in the humanoid robotics sector.

Interesting Points
  • Figure AI achieved production at a rate of 1 robot per hour
  • The company teased plans for a large fleet deployment
  • This represents a major scaling milestone for humanoid robotics manufacturing

Talkie, a 13B LM trained exclusively on pre-1931 data

2473 points · 366 comments · r/singularity · by u/Outside-Iron-8242

A 13-billion parameter language model called Talkie was trained exclusively on data published before 1931, creating a model that speaks and thinks in a distinctly historical register. The project demonstrates how training data cutoffs fundamentally shape a model's worldview, language patterns, and knowledge base.

Interesting Points
  • The model was trained exclusively on pre-1931 data, creating a distinctly historical voice
  • The project demonstrates how training data cutoffs fundamentally shape a model's worldview
  • At 13B parameters, it shows that smaller models can produce compelling results with carefully curated data

Mozilla Used Anthropic's Mythos to Find and Fix 271 Bugs in Firefox

884 points · 111 comments · r/singularity · by u/Tinac4

Mozilla Firefox logo

Mozilla used Anthropic's Claude Mythos model to identify and fix 271 zero-day vulnerabilities in Firefox. The discovery demonstrates the practical cybersecurity applications of frontier AI models, even as the same model's capabilities have sparked debate about whether such tools are too dangerous to release.

Interesting Points
  • Claude Mythos found 271 zero-day vulnerabilities in Firefox
  • The same model that Mozilla praised for its cybersecurity capabilities was described by Anthropic as potentially having 'world-altering consequences' if misused
  • This represents one of the most significant real-world applications of frontier AI for security research

16x DGX Sparks - What should I run?

1107 points · 494 comments · r/LocalLLaMA · by u/Kurcide

DGX Spark hardware

A community member with access to 16x DGX Spark systems asks the r/LocalLLaMA community for recommendations on what models and workloads to run. The post generated extensive discussion about optimal model sizes, quantization strategies, and practical use cases for the hardware configuration.

Interesting Points
  • The poster has access to 16x DGX Spark systems, representing significant local inference compute
  • The post generated 494 comments with extensive community discussion on model selection and quantization
  • Reflects the growing accessibility of powerful local inference hardware

Mistral Medium 3.5 128B is launched

479 points · 281 comments · r/LocalLLaMA · by u/jacek2023

Mistral Medium 3.5 model card

Mistral AI released Mistral Medium 3.5, a dense 128B parameter model with a 256k context window. It is Mistral's first flagship merged model, handling instruction-following, reasoning, and coding in a single set of weights. The model supports configurable reasoning effort per request, multimodal input with a vision encoder trained from scratch, and is released under a Modified MIT License.

Interesting Points
  • 128B dense model with 256k context window, replacing both Mistral Medium 3.1 and Magistral
  • Configurable reasoning effort allows the same model to handle quick replies or complex agentic runs
  • Released under a Modified MIT License for both commercial and non-commercial use

GPT-6 Confirmed

844 points · 59 comments · r/OpenAI · by u/DigSignificant1419

GPT-6 announcement

A post confirming the existence of GPT-6 generated significant discussion in the r/OpenAI community. The confirmation comes amid broader industry shifts including GPT-5.5 landing on AWS and OpenAI's ongoing legal and organizational challenges.

Interesting Points
  • GPT-6 has been confirmed, adding to the rapidly evolving model landscape
  • The post generated 59 comments discussing implications for the AI industry
  • Comes amid GPT-5.5's recent availability on AWS, ending Microsoft's seven-year exclusive partnership

Ai is getting too realistic

3686 points · 588 comments · r/ChatGPT · by u/Remarkable-Sir4051

AI-generated realistic image

A viral post expressing concern that AI-generated content is becoming increasingly realistic, sparking widespread discussion about the implications for media authenticity, deepfakes, and public trust in digital content.

Interesting Points
  • The post received 3686 points and 588 comments, indicating strong community engagement
  • Reflects growing public anxiety about AI-generated media becoming indistinguishable from reality
  • The discussion touches on broader concerns about deepfakes and information integrity

The cost of compute is far beyond the costs of the employees: Nvidia exec says right now AI is more expensive than paying human workers

438 points · 123 comments · r/artificial · by u/chunmunsingh

Nvidia executive discussing AI costs

An Nvidia executive stated that the cost of AI compute infrastructure currently far exceeds the cost of human workers, suggesting that AI deployment may not yet be economically viable for many use cases. This contradicts the narrative that AI is already cheaper than human labor and adds to the ongoing debate about AI economics.

Interesting Points
  • Nvidia exec stated AI compute costs are currently higher than paying human workers
  • The comment adds to the broader debate about whether AI business models are sustainable
  • Contradicts the common narrative that AI is already cheaper than human labor

Dall E 3 vs Image 2.0

790 points · 81 comments · r/OpenAI · by u/RealMelonBread

DALL-E 3 vs Image 2.0 comparison

A comparison post between DALL-E 3 and OpenAI's newer Image 2.0 model generated significant discussion in the r/OpenAI community about the quality differences, use cases, and practical improvements between the two image generation systems.

Interesting Points
  • The post compares DALL-E 3 with OpenAI's newer Image 2.0 model
  • Generated 81 comments discussing quality differences and practical improvements
  • Reflects ongoing community interest in comparing OpenAI's image generation capabilities

Google just released Deep Research Max — an autonomous research agent that writes expert-grade reports on its own

76 points · 30 comments · r/artificial · by u/demchaav

Google released Deep Research Max, an autonomous research agent built on Gemini 3.1 Pro that can search the web, reason over sources, and produce fully cited professional-grade reports with native charts and infographics. Two modes are available: Deep Research for faster real-time responses and Deep Research Max for more thorough analysis.

Interesting Points
  • Built on Gemini 3.1 Pro with MCP integration for private data access
  • Produces fully cited reports with native charts and infographics
  • Two modes: faster real-time responses and more thorough Deep Research Max analysis

Quick Mentions

Report generated in 3m 12s.