· 11:55 PM PDT

DeepSeek V4 and GPT-5.5 Reshape the AI Frontier

Overview

The AI landscape is shifting rapidly as DeepSeek releases its massive open-weight V4 models and OpenAI rolls out GPT-5.5 with aggressive agentic capabilities. Simultaneously, debates over hosted model degradation, AI cybersecurity breakthroughs, and the decoupling of capability from moral accountability dominate community discussions.


Hacker News Stories

DeepSeek v4

1877 points · 1459 comments · by impact_sy

DeepSeek V4 announcement banner

DeepSeek has released V4-Pro and V4-Flash, open-sourcing frontier-level models that support 1 million token context windows. V4-Pro features 1.6 trillion total parameters with 49 billion active, delivering agentic coding and reasoning capabilities that rival top closed-source models while offering API pricing at a fraction of the cost.

Interesting Points
  • V4-Pro has 1.6T total / 49B active params
  • V4-Flash has 284B total / 13B active params
  • Supports 1M context as standard across all official services
  • Pro pricing is $3.48 / 1M output tokens
  • Flash pricing is $0.28 / 1M output tokens
Top Comment Threads
  1. jari_mustonen (23 replies) -- Praises the open-source approach and developer documentation, noting the zero CUDA dependency and complete Chinese AI stack running on Huawei chips as a major shift that breaks monopolies.
  2. revolvingthrow (18 replies) -- Questions the narrative that frontier labs are bleeding money on inference, pointing out that V4 Pro's $3.48/1M output price suggests even the API is highly profitable, with subscriptions likely covering costs.
  3. sudo_cowsay (9 replies) -- Raises security concerns about using Chinese LLMs, prompting discussion on supply chain attacks, training data poisoning, and whether Western models pose similar or different risks.

GPT-5.5

1539 points · 1025 comments · by rd

OpenAI has released GPT-5.5, positioning it as a major step toward a unified AI super app that combines ChatGPT, coding tools, and browser capabilities. The model delivers significant improvements in agentic coding and intuitive task handling, though some developers report issues with model laziness and over-optimization for token efficiency.

Interesting Points
  • Described as OpenAI's 'smartest and most intuitive to use model' yet
  • Focused on agentic coding and real work tasks
  • Part of OpenAI's vision for a unified AI super app
  • Rollout in ChatGPT and Codex is gradual to maintain stability
Top Comment Threads
  1. alternator (40 replies) -- Expresses unease about the deep dependency on frontier coding models, comparing it to playing a game on God Mode. Notes that engineers are becoming so reliant that losing access feels like an amputation, raising concerns about skill atrophy.
  2. noosphr (12 replies) -- Argues that LLMs upend traditional labor theory by turning intelligence into a form of capital that can be withheld indefinitely. Warns that companies not running in-house models are signing up for provider leverage.
  3. aliljet (16 replies) -- Asks the community how to avoid vendor lock-in with Claude or OpenAI ecosystems. Suggestions include using generic agent files, symlinking workflows, and treating official clients as dumb inference gateways.

OpenAI releases GPT-5.5 and GPT-5.5 Pro in the API

232 points · 126 comments · by arabicalories

OpenAI API changelog banner

GPT-5.5 and GPT-5.5 Pro are now available via the OpenAI API, with pricing set at $30 per 1 million output tokens. The API documentation lists a knowledge cutoff of December 1, 2025, though prompting reveals the model itself claims a June 2024 cutoff.

Interesting Points
  • GPT-5.5 Pro is available alongside the standard GPT-5.5 model
  • API pricing is $30 / 1M output tokens
  • API metadata lists a Dec 1, 2025 knowledge cutoff
  • Model self-reports a June 2024 knowledge cutoff when prompted
Top Comment Threads
  1. wincy (5 replies) -- Reports that GPT-5.5 can be lazy, sometimes refusing to fully execute instructions like writing a complete SQL transaction and instead telling the user to do it themselves to save tokens.
  2. czk (4 replies) -- Notices a discrepancy between the API page stating a Dec 2025 knowledge cutoff and the model's own system prompt claiming June 2024. Speculates OpenAI may intentionally feed an older date to encourage tool calls and web searches for tuning.
  3. guilamu (4 replies) -- Shares a self-hosted WordPress benchmark where GPT-5.5 performed poorly compared to other models, questioning how a top-tier model can underperform on practical coding tasks.

GPT-5.5: Mythos-Like Hacking, Open to All

75 points · 23 comments · by rs_rs_rs_rs_rs

XBOW blog post about GPT-5.5 cybersecurity

Security firm XBOW reports that GPT-5.5 delivers a step change in vulnerability detection comparable to Anthropic's exclusive Mythos model. Tested inside their penetration testing agent workflows, GPT-5.5 achieved the best performance on their internal benchmarks for finding and exploiting known open-source vulnerabilities.

Interesting Points
  • GPT-5.5 outperformed previous models on XBOW's internal vulnerability detection benchmarks
  • The model was evaluated inside real agent workflows, not in isolation
  • Focuses on blackbox and whitebox security testing across OSS applications
  • Marks a shift where top-tier hacking capabilities are becoming widely accessible
Top Comment Threads
  1. nsingh2 (2 replies) -- Critiques the blog's data visualization, noting that connecting categorical data with lines implies non-existent data points and makes the plots look like an thinly disguised advertisement.
  2. strange_quark (2 replies) -- Questions how this differs from small open-weight models, citing prior reports that they could detect most headline vulnerabilities. Another commenter counters that open models struggle with false positives and cannot autonomously build and run exploits like Mythos was claimed to do.
  3. JellyYelly (1 replies) -- Argues that to be truly Mythos-like, a model must produce as many novel, high-severity security vulns outlined in Anthropic's redteam blog. Notes that Anthropic has produced no public proof of their own capabilities.

Unauthorized Discord group gained access to Anthropic's Mythos model

9 points · 1 comments · by thoughtpeddler

Reports claim that an unauthorized Discord group managed to gain access to Anthropic's highly restricted Mythos model, an AI tool described as 'too dangerous to release.' The breach highlights security vulnerabilities surrounding exclusive, frontier-level AI cybersecurity capabilities.

Interesting Points
  • Mythos was restricted to a select group of partners for early access
  • Described by Anthropic as an exclusive cybersecurity tool 'too dangerous to release'
  • An unauthorized Discord group reportedly accessed the model
  • Raises questions about the security of exclusive, frontier AI deployments
Top Comment Threads
  1. gnabgib (0 replies) -- Links to a smaller HN discussion from the previous day about the Mythos announcement, noting the community was already tracking the story.

Reddit Stories

DeepSeek v4 people

1949 points · 272 comments · r/LocalLLaMA · by u/markeus101

Community reaction image

The LocalLLaMA community erupts in discussion over DeepSeek's V4 release, with users sharing benchmarks, testing local inference, and debating the geopolitical and economic implications of China's open-weight models competing directly with US closed-source leaders.

Interesting Points
  • DeepSeek V4 is widely discussed as a milestone for open-weight AI
  • Community members share local inference setups and quantization strategies
  • Debate centers on whether Chinese models represent healthy competition or geopolitical risks
Top Comment Threads
  1. u/redditscraperbot2 (1286 points · permalink) -- Notes that the model's ability to recognize classic riddles confirms it's heavily trained on existing internet data, making the 'AI sentience' hype outdated.
  2. u/Automatic-Arm8153 (439 points · permalink) -- Validates the community's suspicion that hosted models were deliberately downgraded, noting that Anthropic's admission proves the value of open weights for stable performance.
  3. u/dwrz (118 points · permalink) -- States that if hosted models are reduced in capability, the price should drop accordingly. Praises local tools like llama.cpp for providing consistent performance without corporate throttling.

Anthropic admits to have made hosted models more stupid, proving the importance of open weight, local models

1111 points · 226 comments · r/LocalLLaMA · by u/spaceman_

Article screenshot about Anthropic's admission

Following Anthropic's acknowledgment that they reduced default thinking/reasoning steps in hosted Claude models to optimize token spend and profit margins, the community praises open-weight models for offering consistent, unthrottled performance.

Interesting Points
  • Anthropic admitted to reducing default reasoning steps to cut token spend
  • Hosted models are optimized for profit rather than raw capability
  • Open-weight models provide a reliable alternative不受 subscription limit changes
Top Comment Threads
  1. u/Automatic-Arm8153 (439 points · permalink) -- Validates the community's suspicion that hosted models were deliberately downgraded, noting that Anthropic's admission proves the value of open weights for stable performance.
  2. u/dwrz (118 points · permalink) -- States that if hosted models are reduced in capability, the price should drop accordingly. Praises local tools like llama.cpp for providing consistent performance without corporate throttling.
  3. u/Important-Radish-722 (90 points · permalink) -- Jokes that lowering model quality forces users to ask more questions, burning more tokens, which ironically benefits the AI companies' revenue streams.

Exactly 1 year ago, Anthropic said fully AI employees were just 1 year away

1057 points · 234 comments · r/singularity · by u/Distinct-Question-16

Image related to AI employees prediction

A retrospective on Anthropic's past predictions regarding fully autonomous AI employees sparks debate about the current state of AI agents. The community discusses whether the timeline was optimistic or if cost and reliability barriers are the real constraints.

Interesting Points
  • Anthropic previously predicted fully AI employees would be available within a year
  • Community debate focuses on whether the delay is due to technology limits or cost
  • Discussion highlights the gap between corporate predictions and practical deployment
Top Comment Threads
  1. u/GrapefruitMammoth626 (185 points · permalink) -- Argues that with a good harness and unlimited tokens, AI employees are theoretically possible today, but economic displacement and cost prevent widespread adoption.
  2. u/ConfidentReality9024 (119 points · permalink) -- Believes the timeline was off but warns that the breakthrough will arrive when people least expect it.
  3. u/stellar_opossum (71 points · permalink) -- Skeptical of the prediction, noting that similar past claims have consistently missed the mark and that AI is still far from replacing human employees reliably.

Mozilla Used Anthropic’s Mythos to Find and Fix 271 Bugs in Firefox

858 points · 105 comments · r/singularity · by u/Tinac4

Firefox and Mozilla logo

Mozilla announced that the upcoming Firefox 150 release includes protections for 271 vulnerabilities identified using early access to Anthropic's exclusive Mythos Preview model. A Mozilla employee clarifies that these internally found bugs are grouped into roll-up advisories rather than individual CVEs.

Interesting Points
  • Firefox 150 patches 271 vulnerabilities found by Anthropic's Mythos model
  • Mythos was used in a closed partnership to proactively hunt bugs
  • Mozilla groups internally found bugs into roll-up security advisories
  • Highlights practical, immediate applications of frontier AI in software security
Top Comment Threads
  1. u/EvillNooB (324 points · permalink) -- Asks how to get access to Mythos, jokingly hoping it will fix their life. Another user notes that Anthropic is distributing it to companies to prepare for upcoming cyber threats.
  2. u/helg0ret (84 points · permalink) -- Questions why Firefox 150's changelog only mentions 3 CVEs. A Mozilla employee explains that internally found bugs are consolidated into roll-up advisories to avoid spamming the public CVE list.
  3. u/benl5442 (50 points · permalink) -- Predicts that AI-driven security will eventually lead to nightly security releases, as vulnerabilities can be exploited instantly once found.

A Yale ethicist who has studied AI for 25 years says the real danger isn’t superintelligence. It’s the absence of moral intelligence.

284 points · 99 comments · r/artificial · by u/reesefinchjh

Yale ethicist interview

Yale ethicist Wendell Wallach argues that the primary risk of AI is not the emergence of superintelligence, but the dangerous decoupling of technical capability from moral reasoning and accountability. He warns that as systems become more autonomous, liability gets diffused across developers, users, and corporations.

Interesting Points
  • Yale ethicist Wendell Wallach warns that capability is scaling faster than moral reasoning
  • AI systems can be extraordinarily intelligent while having zero moral judgment
  • The 'accountability gap' is identified as the real black swan risk of 2026
  • Liability is diffused across developers, users, and corporations, automating away responsibility
Top Comment Threads
  1. u/IsThisStillAIIs2 (17 points · permalink) -- Agrees that capability is outpacing boundaries, emphasizing the accountability problem where harm is diffused across a complex system of developers, users, and corporations.
  2. u/ItsAConspiracy (3 points · permalink) -- Questions whether an ASI would actually be dangerous if it lacks moral reasoning, pointing out that high capability without alignment is a classic safety concern.
  3. u/Shot_Ideal1897 (3 points · permalink) -- Frames the decoupling of capability from consequence as the real failure in how we define agency. Notes that developers, users, and corporations all point fingers when systems cause systemic harm.

Quick Mentions

Report generated in 18m 39s.