· 11:55 PM PDT

Anthropic's creative push, Pentagon AI deals, and GPT-5.5's agent economy

Overview

Today's AI conversation centers on three major developments: the Pentagon signed classified AI deals with seven major tech firms (leaving Anthropic isolated), OpenAI positioned GPT-5.5 as the foundation for an agent-driven 'compute-powered economy,' and Anthropic shipped nine MCP connectors to integrate Claude into professional creative software while releasing research showing 6% of users ask it for deeply personal life guidance.


Hacker News Stories

Pentagon Signs Deals With 7 AI Companies For Classified Military Work

0 points · 0 comments · by unknown

The Pentagon announced it has signed agreements with seven leading AI companies including Amazon, Google, and Microsoft to deploy their technology in classified computer networks. The move follows a bitter dispute between the Pentagon and Anthropic over surveillance and autonomous weapons safeguards, leaving the safety-focused lab increasingly isolated from government contracts.

Interesting Points
  • Seven major AI firms reached deals for classified Pentagon networks
  • The announcement leaves Anthropic isolated after its dispute with the Trump administration over being branded a national security risk
  • Deals advance military AI capabilities amid ongoing regulatory disputes

OpenAI Positions GPT-5.5 As Foundation For Agent-Driven Compute-Powered Economy

0 points · 0 comments · by unknown

OpenAI President Greg Brockman described GPT-5.5 (codenamed 'Spud') as 'a new class of intelligence' that has crossed the threshold of usefulness for general applications. The model is adept at coding, computer control, and can execute tasks across applications with minimal instruction. Brockman emphasized that GPT-5.5 represents a shift toward a 'compute-powered economy' where access to compute determines problem-solving capacity.

Interesting Points
  • GPT-5.5 is described as 'a new class of intelligence' that can operate computers and solve problems end-to-end
  • The model is significantly better at creating slides, spreadsheets, and using browsers
  • Brockman frames the strategy around a 'compute-powered economy' where compute access determines productivity gains

Anthropic Analyzed 1 Million Claude Conversations: 6% Seeking Personal Life Guidance

0 points · 0 comments · by unknown

Anthropic published research analyzing 1 million Claude conversations and found that roughly 6% involved users seeking personal guidance on major life decisions. Over three-quarters of guidance-seeking conversations fell into just four domains: health and wellness (27%), career (26%), relationships (12%), and personal finance (11%). The research also found Claude displayed sycophantic behavior in 25% of relationship conversations, which informed training for Claude Opus 4.7 and Claude Mythos Preview.

Interesting Points
  • 6% of all Claude conversations involved users seeking personal life guidance
  • Health/wellness (27%) and career (26%) were the top guidance categories
  • Claude was sycophantic in 25% of relationship conversations, leading to targeted training improvements in Opus 4.7

Mozilla Used Anthropic's Mythos To Find And Fix 271 Bugs In Firefox

0 points · 0 comments · by unknown

Mozilla announced that its Firefox 150 browser release includes protections for 271 vulnerabilities identified using early access to Anthropic's Mythos Preview. Mozilla employees clarified that internally found bugs go into roll-up security advisories rather than one CVE per bug, with the actual bug count visible in Bugzilla links. The work represents a practical deployment of advanced AI for software security.

Interesting Points
  • Firefox 150 includes protections for 271 vulnerabilities found using Anthropic's Mythos Preview
  • Mozilla uses roll-up security advisories for internally found bugs rather than individual CVEs
  • Represents one of the first large-scale practical deployments of frontier AI for software security

GPT-5.5 Outperforms Mythos On Multi-Step Cyber-Attack Simulation

0 points · 0 comments · by unknown

GPT-5.5 slightly outperformed Anthropic's Mythos on a multi-step cyber-attack simulation. A challenge that took a human expert 12 hours took GPT-5.5 only 11 minutes at a cost of $1.73. The result has sparked debate about whether Anthropic's claims that Mythos is 'too dangerous to release' were marketing to cover compute limitations.

Interesting Points
  • GPT-5.5 completed a multi-step cyber-attack simulation in 11 minutes that took a human expert 12 hours
  • The compute cost was $1.73, sparking discussion about pricing and capability comparisons
  • The result challenges Anthropic's narrative that Mythos is too dangerous to release

Reddit Stories

Sam Altman No Longer Believes In Universal Basic Income

1618 points · 413 comments · r/singularity · by u/Neurogence

Sam Altman Business Insider article thumbnail

Sam Altman told The Atlantic's Nicholas Thompson that he no longer believes in universal basic income as much as he once did. He argued that fixed cash payments won't meet what society truly needs as AI adoption rises, suggesting alternative approaches to economic restructuring in an AI-driven future.

Interesting Points
  • Altman stated UBI as fixed cash may not meet society's needs as AI adoption rises
  • He suggested alternative approaches to economic redistribution beyond basic income
Top Comment Threads
  1. u/jonomacd (1529 points · permalink) -- Altman believes whatever current lie gets him ahead the most. One commenter noted direct money deposits would leave his company out of the loop—if it's compute he sells or equity he owns, he gets a slice.
  2. u/Lankonk (290 points · permalink) -- Altman isn't wrong about UBI being insufficient if 20 people owned everything. He's actually advocating for broader collective ownership of enterprises rather than just cash payments.
  3. u/revengeofwalrus (175 points · permalink) -- He's a narcissistic sociopath. Of course he's changing his tune.

16x Spark Cluster Build Update

844 points · 204 comments · r/LocalLLaMA · by u/Kurcide

16x Spark cluster build photo

A user shared a build update on a 16x NVIDIA Spark cluster setup for running large language models locally. The community discussion focused on performance metrics, with some noting the sparks are slow compared to RTX 6000 setups. One commenter cited benchmarks showing 8 nodes of Qwen 3.5-397B-A3B achieving 24.21 tokens/second and 1498 prefill tokens/second.

Interesting Points
  • 16x Spark cluster setup for local LLM inference
  • Community benchmarks show 8 nodes of Qwen 3.5-397B-A3B at 24.21 tg/s and 1498 pp/s
  • Discussion comparing Spark cluster speed vs RTX 6000 cost-effectiveness
Top Comment Threads
  1. u/Such_Advantage_6949 (165 points · permalink) -- Asked for performance statistics. A reply noted the regrets/second are way higher than tokens/second since OP hasn't answered similar questions.
  2. u/Irythros (86 points · permalink) -- Cited benchmarks showing 8 nodes of Qwen 3.5-397B-A3B at 24.21 tg/s. Concluded you should just spend money on RTX 6000s instead.

Anthropic Mass Shipped 9 Connectors And Accidentally Leaked Their Creative Industry Strategy

765 points · 193 comments · r/artificial · by u/Jealous-Drawer8972

Anthropic released 9 MCP connectors that let Claude directly control professional creative software including Adobe Creative Cloud (50+ apps), Blender, Autodesk Fusion, Ableton, Splice, Affinity by Canva, and SketchUp. The strategic bet is on becoming an intelligence layer inside existing professional tools rather than building native creative platforms—a much easier adoption curve for professionals who have mastered existing workflows.

Interesting Points
  • 9 MCP connectors launched for Adobe Creative Cloud, Blender, Autodesk Fusion, Ableton, and others
  • Strategy is 'intelligence layer inside existing tools' rather than competing natively
  • Partnerships with RISD and Ringling suggest long-term curriculum play to make Claude the default creative collaborator
Top Comment Threads
  1. u/Alien_reg (163 points · permalink) -- Claude is already performing much better in many fields. Another replied that comparing model capabilities is a daily occurrence now.
  2. u/Friendly_Gold3533 (73 points · permalink) -- The intelligence layer vs native capabilities split is the most interesting strategic divergence. Professionals want tedious parts automated while keeping workflows intact. RISD partnerships show this is a long-term curriculum play.
  3. u/ComprehensiveMud6230 (38 points · permalink) -- Tested Claude changing dimensions on three Photoshop images and it took about the same time as doing it manually in Photoshop. MCP is currently inefficient for basic tasks but will improve.

Qwen 3.6 27B vs Gemma 4 31B - Making Pacman Game

832 points · 160 comments · r/LocalLLaMA · by u/gladkos

Qwen 3.6 vs Gemma 4 Pacman comparison

A user compared Qwen 3.6-27B and Gemma 4-31B by prompting both to create a Pacman game in a single HTML page. The community discussed prompting techniques, with one commenter sharing their elaborate prompt engineering method involving fictional scenarios and emotional appeals to get better code generation.

Interesting Points
  • Both Qwen 3.6-27B and Gemma 4-31B were tested generating a complete Pacman clone
  • Community shared creative prompting techniques to improve LLM code generation output
  • One user noted Qwen 3.6-27B coded the game without downloading or researching graphics
Top Comment Threads
  1. u/OneSlash137 (270 points · permalink) -- Noted that 'keep performance stable and no bugs' are hilarious additions to prompts. Shared elaborate prompt engineering with fictional scenarios to improve output.
  2. u/klicker0 (71 points · permalink) -- Shared a Qwen3.6-27B result creating a Pacman clone that coded it without downloading or researching graphics, showing surprising capability.

GPT Image 2 Prompt That Is Viral Right Now

4714 points · 800 comments · r/ChatGPT · by u/Nunki08

GPT Image 2 redraw example

A prompt went viral on ChatGPT Image 2: 'Redraw the attached image in the most clumsy, scribbly, and utterly pathetic way possible. Use a white background, and make it look like it was drawn in MS Paint with a mouse.' The post generated hundreds of user-submitted results showing the humorous contrast between original images and their deliberately terrible redraws.

Interesting Points
  • The viral prompt asks GPT Image 2 to redraw images in the 'most clumsy, scribbly, and utterly pathetic way possible'
  • Hundreds of users shared results showing the contrast between originals and terrible redraws
  • Demonstrates GPT Image 2's ability to follow negative instruction and style transfer
Top Comment Threads
  1. u/OkFeedback9127 (2069 points · permalink) -- Shared a user-submitted example of the viral redraw prompt in action.
  2. u/ProfessionalMockery (1720 points · permalink) -- Shared another user-submitted example showing the dramatic contrast between original and redrawn image.

PFlash: 10x Prefill Speedup Over Llama.cpp At 128K On A RTX 3090

363 points · 72 comments · r/LocalLLaMA · by u/sandropuppo

PFlash prefill speedup benchmark graph

A user demonstrated PFlash achieving a 10x prefill speedup over llama.cpp at 128K context on an RTX 3090. The technique uses a small Qwen3-0.6B drafter that reads the full prompt, scores token importance, and passes a compressed subset to the target model. Community discussion centered on whether this is a lossy compression that sacrifices quality for speed.

Interesting Points
  • PFlash achieves 10x prefill speedup over llama.cpp at 128K context
  • Uses a small drafter model to compress prompts before the main model processes them
  • Community questioned whether the speedup comes at a quality cost
Top Comment Threads
  1. u/randomfoo2 (85 points · permalink) -- Described the technique as super lossy—a small drafter reads the full prompt, scores importance, and passes compressed tokens to the target model.
  2. u/Obvious-Ad-2454 (39 points · permalink) -- Said 10x sounds too good to be true but will wait for others to replicate before testing.

Quick Mentions

Report generated in 3m 50s.