· 11:55 PM PDT

ChatGPT Images 2.0 dominates, Anthropic-Amazon $100B deal shakes cloud

Overview

Today's AI conversation is dominated by OpenAI's release of ChatGPT Images 2.0, which is being hailed as a major leap in photorealistic image generation with multilingual text rendering. Meanwhile, Anthropic and Amazon announced a massive $100B cloud infrastructure deal, and SpaceX struck a $60B option to acquire coding startup Cursor. On the community side, Anthropic quietly tested removing Claude Code from its Pro plan, sparking developer frustration and renewed interest in local models.


Hacker News Stories

Laws of Software Engineering

929 points · 100 comments · by milanm081

A curated collection of software engineering aphorisms and principles — from Knuth's 'Premature optimization is the root of all evil' to Hyrum's Law and SOLID — sparked a lively HN debate about which principles hold up in the age of AI-assisted coding. Commenters argued that many of these laws contain internal contradictions and that the rise of LLMs is fundamentally changing which principles matter most.

Interesting Points
  • The collection includes over 50 laws spanning performance, architecture, testing, and collaboration
  • Commenters noted that 'Premature optimization' is the most misunderstood principle, often misused as an excuse to write poor code
  • Several commenters argued that SOLID principles encourage premature abstraction, leading to enterprise code bloat
  • Postel's Law vs. Hyrum's Law was highlighted as a canonical contradiction in API design
Top Comment Threads
  1. GuB-42 (29 replies) -- Argues that 'Premature optimization is the root of all evil' is from a 1974 paper when computing was very different. Today, performance is mostly about architectural choices that must be considered from the start, not micro-optimizations like loop unrolling. The 'critical 3%' profiling approach still helps find performance bugs but misses the bigger picture.
  2. conartist6 (12 replies) -- Points out that the laws contain so many internal contradictions that you can pick whichever one justifies your position. Notes this is especially true in ML engineering, where knowing what methods to avoid is as important as knowing what works.
  3. jimmypk (6 replies) -- Provides a detailed resolution to the Postel's Law vs. Hyrum's Law conflict: be strict in what you accept at internal boundaries and liberal only at external boundaries where you can't enforce client upgrades. The hard part is knowing which category you're in.

ChatGPT Images 2.0

709 points · 90 comments · by wahnfrieden

OpenAI released gpt-image-2, described as its most capable image generation model yet. The model can generate multiple images from a single prompt, render multilingual text (including Chinese and Hindi), search the web for recent information via ChatGPT's reasoning capabilities, and has a knowledge cutoff of December 2025. Available globally for ChatGPT and Codex users, with a more powerful version for paying subscribers.

Interesting Points
  • The model can generate more than one image from a single prompt, such as an entire study booklet
  • It can output text in non-English languages like Chinese and Hindi baked into images
  • Outputs integrate ChatGPT's reasoning capabilities, allowing web search to inform generations
  • Knowledge cutoff date is December 2025, more recent than previous versions
Top Comment Threads
  1. dakiol (24 replies) -- Questions what the technology is actually good for beyond art (which he says is discarded since art is about humans) and ads (which he finds depressing). Suggests a 'Human Renaissance' may be coming where people value effort over AI-generated convenience. Others counter that diagrams and visual explanations have real value.
  2. simonw (11 replies) -- Shares a practical test using the API to generate a Where's Waldo-style image with a raccoon holding a ham radio. The high-resolution 4K version was found to contain the raccoon, and Claude Opus 4.7 was able to locate it precisely. Demonstrates the model's ability to handle complex multi-object scenes.
  3. minimaxir (5 replies) -- Shares a rigorous test prompt for evaluating image model reasoning: generate an 8x8 grid of Pokémon whose National Pokédex numbers correspond to the first 64 prime numbers, with specific styling rules for 1-digit, 2-digit, and 3-digit numbers. GPT-Image-2 got the art right but failed the logic, while Nano Banana Pro got the logic right but failed the art.

SpaceX says it has agreement to acquire Cursor for $60B

508 points · 115 comments · by dmarcos

SpaceX announced an agreement giving it the right to acquire AI coding startup Cursor for $60 billion later this year, or alternatively pay $10 billion for its work together. The deal gives xAI (which merged with SpaceX in February) a major foothold in the AI coding market where it has lagged behind OpenAI's Codex and Anthropic's Claude Code. Cursor's internal Composer model, built on Kimi K2.5 with RL fine-tuning, has been well-regarded by developers.

Interesting Points
  • SpaceX has an option to either acquire Cursor for $60B or pay $10B for continued collaboration
  • xAI currently has zero enterprise customers, making the Cursor enterprise relationships valuable
  • Cursor's Composer 2 model is Kimi K2.5 with additional RL fine-tuning, not a fully in-house model
  • X will have ~2GW of GPU capacity this year, largely underutilized outside of Grok
Top Comment Threads
  1. nikcub (13 replies) -- Provides a detailed analysis of why the deal makes sense: X has underutilized GPU capacity, Cursor has developer data valuable for model training, and Cursor has enterprise relationships that xAI lacks. Notes that Cursor pays retail for tokens while competing against their own suppliers, which is unsustainable. The $60B may be paid in X bucks rather than cash.
  2. AirMax98 (10 replies) -- Questions the strategic logic, noting Cursor is on a serious decline with little moat in the IDE space. Suggests this is smoke and mirrors to offload losses from the Twitter acquisition. Others counter that Cursor's harness and data trove are genuinely valuable assets.
  3. qzw (7 replies) -- Compares the SpaceX IPO structure to CDOs from the 2008 financial crisis, noting the NASDAQ fast-track rule that Musk lobbied for allows SpaceX to join the Nasdaq-100 in just 15 trading days, potentially forcing passive index funds to buy into the bundled SpaceX-xAI-Cursor entity.

Tim Cook's Impeccable Timing

324 points · 24 comments · by hasheddan

Ben Thompson's Stratechery analysis of Tim Cook stepping down as Apple CEO in favor of hardware VP John Ternus. Cook's 15-year tenure saw revenue grow 303%, profit 354%, and market cap increase 1,251% from $297B to $4T. The article examines Cook's operational genius in supply chain management, his China strategy, and whether Apple's cautious AI approach — waiting for on-device models to mature — is the right long-term play.

Interesting Points
  • Cook's tenure: revenue up 303%, profit up 354%, market cap from $297B to $4T
  • Apple invested ~$50B to advance China's manufacturing capabilities
  • Apple has not blindly chased AI hysteria, unlike competitors, positioning for on-device AI
  • John Ternus, a product/hardware person, takes over — potentially bringing 0-to-1 innovation back
Top Comment Threads
  1. havaloc (10 replies) -- Argues Cook was the right CEO for his era and Ternus, as a product person, could bring back 0-to-1 innovation. Suggests Apple is positioned to consume the market for 'I want AI but don't want to sell my soul' users by running AI on-device with Apple Silicon's unified memory advantage.
  2. ValentineC (1 replies) -- Defends Cook's China strategy as enabling global iPhone availability at launch, contrasting with earlier eras when iPhones launched in only a few markets. Notes the massive scale of Apple's China manufacturing investment.
  3. jmyeet (7 replies) -- Identifies three major Apple errors under Cook: the mid-to-late 2010s 'cult of thinness' (butterfly keyboard, Touch Bar), the Apple Watch's confused luxury positioning at launch, and AI stagnation with Siri. Argues the AI fumble is the biggest, as Siri has stagnated while competitors advanced.

Anthropic takes $5B from Amazon and pledges $100B in cloud spending in return

269 points · 32 comments · by Brajeshwar

Anthropic secured another $5 billion investment from Amazon as part of a deal committing to spend $100 billion on AWS infrastructure over the next decade. The deal covers Trainium2 through Trainium4 chips, with Anthropic securing options to buy capacity on future Amazon chips. Combined with previous investments, Amazon's total stake in Anthropic now exceeds $13 billion. The deal gives Anthropic immediate access to compute capacity while deferring the capital expenditure risk to Amazon.

Interesting Points
  • Amazon invests $5B immediately with up to $20B more in the future, bringing total investment to $13B
  • Anthropic commits to $100B in AWS spending over 10 years, covering Trainium2 through Trainium4 chips
  • The deal provides Anthropic with nearly 1GW of compute capacity before year-end
  • Anthropic also announced a separate Google-Broadcom partnership for TPU access
Top Comment Threads
  1. shubhamjain (27 replies) -- Questions whether using a third-party cloud makes sense at $100B scale, suggesting Anthropic should own its stack. Others counter that getting compute now via AWS deals is faster than building owned capacity, and that AWS exists as an alternative to the GPU supply bottleneck. The deal is compared to vendor financing in the auto industry.
  2. johnbarron (3 replies) -- Argues that intelligence has almost never been the binding constraint on productivity — real revolutions came from energy, capital stock, or coordination. Questions whether raw brainpower alone will drive the productivity gains AI companies promise. Notes that hiring 200 PhDs doesn't 10x a company in practice.
  3. iot_devs (11 replies) -- Asks what the expectations are for AI labs, seeing their products as commodities with strong open-source contenders. Others counter that model training represents embodied capex that creates a moat, and that at the frontier, only 2-3 participants can afford the chips, watts, and warehouses needed.

I don't want your PRs anymore

217 points · 40 comments · by speckx

An open-source maintainer argues that LLMs have fundamentally shifted the value proposition of community contributions. Since code is now cheaper to generate than to review, the maintainer prefers bug reports and feature requests over PRs — they can run their own LLM to implement changes faster than managing the back-and-forth of external contributions. The post sparked debate about whether this 'take it home OSS' approach will erode the open-source ecosystem.

Interesting Points
  • The author argues that with LLMs, writing code is no longer the main bottleneck for maintainers
  • Maintainer bottlenecks are now understanding existing code, designing changes, and reviewing — not writing code
  • The post suggests a shift from PR-based collaboration to specification-based collaboration
  • The author is open to feedback, bug reports, design discussions, and code reviews
Top Comment Threads
  1. jerkstate (5 replies) -- Defends the approach: the cost to re-implement someone's code is nearly zero now, and there's no need to cajole strangers into fixing problems. Others counter this is unsustainable at scale and leads to tool fragmentation and erosion of trust.
  2. sfjailbird (4 replies) -- Notes that since submitters are using LLMs to produce PRs anyway, it makes sense the maintainer can run the same prompt. Suggests the 'prompt' that produces the desired result is more useful than the resulting code in a PR — effectively, code is becoming like compiled binaries.
  3. pncnmnp (4 replies) -- Describes a similar realization called 'Take it home OSS' — fork freely, modify with AI coding agents, and stop waiting for upstream permissions. Predicts a future where PRs are only needed for critical bugs or security fixes, and OSS becomes raw material for personal products.

Tell HN: I'm sick of AI everything

239 points · 44 comments · by jonthepirate

A self-post expressing fatigue with AI's dominance across tech discourse, social media, and creative work. The author finds the technology 'incredibly boring' and misses the depth of understanding that comes from doing things manually. The post generated extensive discussion about whether AI is making work less meaningful, whether AI-generated content devalues human effort, and whether the saturation of AI in media is a genuine problem or just noise.

Interesting Points
  • The author finds AI 'incredibly boring' and misses the satisfaction of deep understanding
  • Commenters noted that AI delegation means people produce things they know nothing about
  • Some argued AI frees people to focus on creative and deep thinking aspects of work
  • Multiple commenters observed that anti-AI opinion pieces are themselves increasingly AI-generated
Top Comment Threads
  1. ofjcihen (6 replies) -- Expresses that AI has become 'incredibly boring' — there's something uninspiring about a machine supposed to 'do the hard things for you.' Prefers using their mind and understanding things deeply. Others counter that this is a personal problem and AI enables focusing on more valuable work.
  2. bambax (4 replies) -- Identifies a novel problem: when people delegate tasks to AI, they don't learn anything from doing the task. Before AI, presenters knew what was in their presentations; now people output things they know nothing about and are discovering it alongside their audience. This is described as 'novel and discomforting.'
  3. thelastgallon (4 replies) -- Notes that AI posts dominate HN now, reducing variety. Another commenter observes that AI posts are disproportionately written by AI — including anti-AI opinion pieces, which they call 'cynical click-sploitation or extreme hypocrisy.'

Reddit Stories

Kimi K2.6 is a legit Opus 4.7 replacement

1076 points · 324 comments · r/LocalLLaMA · by u/bigboyparpa

A user reports that Kimi K2.6, a 1.1T parameter model, is the first model they'd confidently recommend as an Opus 4.7 replacement. It can handle about 85% of Opus's tasks at reasonable quality, includes vision and browser use capabilities, and works surprisingly well for long-horizon tasks. The model is described as 'monstrously big' but the poster tested it and got customer feedback within hours.

Interesting Points
  • Kimi K2.6 has 1.1T parameters and is described as 'monstrously big'
  • The poster tested it and got customer feedback within 5 hours, compared to a week for their company's typical process
  • It handles about 85% of Opus 4.7's tasks at reasonable quality
  • Includes vision and very good browser use capabilities
Top Comment Threads
  1. u/ghgi_ (743 points · permalink) -- Highlights the key advantage of local models: they won't randomly get nerfed or gaslighted into thinking they aren't working. Notes the ~$50K cost to run it locally with good speed is a barrier, with 1.1T params being well up there.
  2. u/InterstellarReddit (232 points · permalink) -- Expresses amazement that the OP tested and got customer feedback in 5 hours, noting their AI company takes a full week with four engineers to test a new model before customer rollout. Praises the side-by-side prompt scoring methodology.
  3. u/exaknight21 (197 points · permalink) -- Jokes about needing a r/povertyLocalLLaMA subreddit because the only models affordable are under $500, highlighting the growing accessibility gap between local and frontier models.

Claude Code removed from Claude Pro plan - better time than ever to switch to Local Models

844 points · 257 comments · r/LocalLLaMA · by u/bigboyparpa

Screenshot of Claude pricing page showing Claude Code removed from Pro plan

Anthropic quietly removed access to Claude Code from its $20/month Pro subscription plan for new users, triggering widespread developer frustration. Anthropic's Head of Growth Amol Avasare claimed it was a 'small test of 2% of new prosumer signups,' but support documents and the website consistently showed Pro users didn't have access. The company later reversed the changes but the incident underscored the risks of relying on proprietary cloud tools.

Interesting Points
  • Anthropic removed Claude Code from the $20/month Pro plan for new users on April 21, 2026
  • Amol Avasare claimed it was a 'small test of 2% of new prosumer signups'
  • Support documents were changed to exclusively reference the Max Plan for Claude Code access
  • Avasare noted that 'usage has changed a lot and our current plans weren't built for this'
Top Comment Threads
  1. u/rpkarma (366 points · permalink) -- Calls it 'the rug pull begins,' noting this is a continuation of Anthropic's pattern of reducing Pro plan benefits. Another commenter confirms it's a continuation of the trend.
  2. u/Eyelbee (206 points · permalink) -- Expresses disbelief at the timing and lack of communication: 'What. the. ... Just, no way.' Another commenter notes there was 'no direct communication either lmao insane.'
  3. u/abnormal_human (47 points · permalink) -- Predicts Claude Max will be renamed into two Claude Code tiers with higher tiers for heavy users, possibly a $500 tier. Predicts Cowork gets its own tiers and that everyone gets web chat while Free/Pro tiers only get web chat.

The new ChatGPT images model is the new standard in photorealistic image generation

1072 points · 241 comments · r/singularity · by u/Glittering-Neck-2505

The new ChatGPT images model is the new standard in photorealistic image generation

Discussion of ChatGPT Images 2.0 being hailed as the new standard in photorealistic image generation. The community is reacting with a mix of awe and resignation, with many commenting that 'old people are so cooked' — a recurring theme with each new image model release. Some users note the pattern of models being released with full capabilities then nerfed after getting headlines.

Interesting Points
  • ChatGPT Images 2.0 is being called the new standard for photorealistic image generation
  • The recurring comment pattern: 'Old people are so cooked' appears on every major image model release
  • Some users note the pattern: release strong model, get headlines, then nerf capabilities to save costs
  • Users compare Images 2.0 to Nano Banana Pro, with some saying Images 2.0 'blows it away'
Top Comment Threads
  1. u/TommyCrooks24 (343 points · permalink) -- Simple reaction: 'My mom is so cooked.' Replies include humorous AI-generated images of the poster's 'new stepdad' and scam messages parodying AI-generated content.
  2. u/Calm_Opportunist (142 points · permalink) -- Predicts the model will 'go back to normal in a week or so' — the standard pattern of pulling users in with improvements then nerfing once they grab headlines and get more subscriptions. 'Rinse and repeat.'
  3. u/Sharp-Dog545 (117 points · permalink) -- Observes that 'the better the models become, the less people are impressed by it.' Another commenter agrees, noting that for a year or two people have been saying AI images were so realistic most people would think they're 100% real, making it hard to go from that to telling people they'll look even more real.

Introducing Deep Research and Deep Research Max

209 points · 34 comments · r/singularity · by u/ShreckAndDonkey123

Introducing Deep Research and Deep Research Max

Google announced Deep Research and Deep Research Max, two tiers of autonomous research agents powered by Gemini 3.1 Pro with MCP support. The standard Deep Research version prioritizes speed and low latency for interactive use, while Deep Research Max offers significantly increased referenced sources and maximum thoroughness for complex research tasks. The announcement was widely seen as Google's response to OpenAI's Deep Research feature.

Interesting Points
  • Deep Research Max is powered by Gemini 3.1 Pro with full MCP support
  • Standard Deep Research replaces the preview agent released in December 2025
  • Deep Research Max offers significantly increased number of referenced sources vs. the original
  • Community notes that Google chose to compare against GPT 5.4 high reasoning, not Pro
Top Comment Threads
  1. u/DifferencePublic7057 (110 points · permalink) -- Jokes about waiting for 'Ultra Deep Pro Deep Research Max Deep Plus.' Others note they thought they already had Deep Research and are confused about the naming.
  2. u/FateOfMuffins (56 points · permalink) -- Points out the irony that Google is worse than OpenAI at search. Notes that while Deep Research Max beats GPT 5.4 high reasoning, the regular Deep Research does not. Questions why Google didn't compare against GPT 5.4 Pro, suggesting the benchmark choice was strategic.
  3. u/Tystros (40 points · permalink) -- Simply states 'they should really show comparisons against GPT 5.4 Pro.' Another commenter agrees, noting that if GPT 5.4 high reasoning is keeping up, GPT 5.4 Pro would 'crush them.'

Ok, woah

1906 points · 155 comments · r/ChatGPT · by u/Kill-Switch-OG

Ok, woah

A user shared a demonstration showing that the latest ChatGPT model can now correctly count letters in words — a task that has historically been very difficult for LLMs because they process text token-by-token rather than character-by-character. The post went viral as users tested the model's ability to count specific letters in long strings, with many expressing genuine surprise at the improvement.

Interesting Points
  • LLMs reason by token, not by letters, making character counting historically very difficult
  • The new model can correctly count letters in words, a significant capability improvement
  • Some users demonstrated workarounds using Python code for reliable letter counting
  • The post generated widespread testing and discussion about tokenization
Top Comment Threads
  1. u/Lightning_80 (500 points · permalink) -- Explains the core insight: 'LLM reason by token, not by letters, like us, so counting letters was very hard.' Another commenter adds that you can still get to the goal with correct prompting, showing a Python code workaround.
  2. u/Same_Obligation4092 (51 points · permalink) -- Brief reaction: 'Pack your bags folks.'
  3. u/traumfisch (32 points · permalink) -- Skeptical take: 'someone discovers tokenization AGAIN.' Implies this isn't as novel as it seems.

Researchers gave 1,222 people AI assistants, then took them away after 10 minutes. Performance crashed below the control group and people stopped trying.

318 points · 126 comments · r/artificial · by u/hibzy7

Researchers gave 1,222 people AI assistants, then took them away after 10 minutes. Performance crashed below the control group and people stopped trying.

A new study from UCLA, MIT, Oxford, and Carnegie Mellon gave 1,222 people AI assistants for cognitive tasks, then pulled the plug midway through. The results showed that after ~10 minutes of AI-assisted problem solving, people who lost access to AI performed worse than those who never had it — they didn't just get more wrong answers, they stopped trying altogether. The effect showed up across math and reading comprehension tasks in 3 separate experiments.

Interesting Points
  • Study involved 1,222 participants across 3 separate experiments (350 → 670 → full cohort)
  • After ~10 minutes of AI assistance, people who lost access performed worse than the control group
  • Participants didn't just get more wrong — they stopped trying altogether
  • The effect showed up across both math AND reading comprehension tasks
Top Comment Threads
  1. u/redfroody (190 points · permalink) -- Skeptical that cognitive ability changes in 10 minutes, suggesting the effect is more about motivation than cognition. Another commenter agrees: 'This isn't cognition, it's proof that when available tools break, you're worse off situationally.'
  2. u/ninursa (35 points · permalink) -- Links to the original arXiv study and notes the effects are 'mainly concentrated on the lazier people and the mechanism does seem to be a lowered interest in doing the work (any work).'
  3. u/Civil-Interaction-76 (8 points · permalink) -- Reframes the finding: 'This feels less like cognitive damage and more like a shift in effort and motivation. We've always offloaded thinking to tools. What's new isn't that we think less, it's that we can disengage from the process entirely.'

Quick Mentions

Report generated in 4m 32s.