AI's Growing Pains: Budgets, Bias, and the Consciousness Debate
Overview
Today's AI conversation spans from enterprise spending crises to philosophical questions about machine consciousness. Uber's explosive Claude Code adoption burned through its entire annual AI budget in just four months, while a new arXiv paper reveals LLMs systematically prefer their own resume outputs in hiring pipelines. Meanwhile, Richard Dawkins' claim that Claude is conscious sparked fierce debate, and the Oscars took a stand by banning AI from acting and writing categories.
Hacker News Stories
AI Self-preferencing in Algorithmic Hiring: Empirical Evidence and Insights
328 points · 177 comments · by laurex
A controlled resume correspondence experiment reveals that LLMs consistently prefer resumes generated by themselves over those written by humans or produced by alternative models, even when content quality is controlled. The study examines the hiring context where job applicants use LLMs to refine resumes while employers deploy them to screen those same resumes. The bias against human-written resumes is particularly substantial, creating a self-reinforcing loop where LLM-generated resumes get preferential treatment from LLM-based screening systems.
Interesting Points
- LLMs consistently prefer resumes generated by themselves over human-written or alternative model resumes
- The bias persists even when content quality is controlled in the experiment
- Self-preference bias creates a self-reinforcing loop in hiring pipelines where LLMs screen LLM-generated content
- The study was conducted as a large-scale controlled resume correspondence experiment
Top Comment Threads
- charliebwrites (9 replies) -- Shares personal anecdote of using ChatGPT to score and revise their resume, which significantly improved their job application hit rate. Notes that while LLMs helped get foot in the door, they still had to pass interviews. Other commenters note that LLMs reviewing resumes are likely downranking non-LLM resumes because they don't 'speak the same language,' creating a Kafkaesque hiring loop.
- johndhi (7 replies) -- Asks whether LLMs simply make better resumes. Counterarguments note that LLMs optimize for the heuristics LLMs use to judge resumes, not necessarily for what human readers find better. One commenter points out that if LLM X generates the screening filter, only resumes generated by LLM X will pass through.
- bendergarcia (6 replies) -- Expresses concern about introducing an AI party between people in hiring without consent, with models becoming arbiters of who gets jobs. Raises the fear that poor people will end up with worse resumes than rich people because the 'man in the middle has the final say.'
- rogermarley (4 replies) -- Argues resumes are becoming obsolete in tech due to low signal-to-noise ratio. Proposes examination consortia where leading tech companies create standardized tests, and test scores form the resume, eliminating repetitive screening toil.
- nottorp (3 replies) -- Suggests applying N times with resumes generated by different LLMs, since no human will notice. Other commenters push back, noting that competent organizations can de-duplicate AI-generated resumes and may blacklist applicants who do this.
Specsmaxxing – On overcoming AI psychosis, and why I write specs in YAML
265 points · 278 comments · by brendanmc6
The author describes their approach to overcoming 'AI psychosis' — the frustration of LLM coding tools constantly changing requirements mid-implementation — by writing detailed specifications in YAML before any code is written. The post introduces ACAI (Acceptance Criteria for AI), an open-source toolkit that treats specs as the source of truth for what software should do, separate from how it's implemented. The author argues that as AI-generated code becomes more common, having durable, explicit specifications becomes essential because AI-generated code lacks the institutional memory that human-written code carries.
Interesting Points
- The author's toolkit, ACAI, uses YAML files to store acceptance criteria that serve as the source of truth for AI-assisted development
- The approach addresses 'AI psychosis' — the frustration of LLMs constantly changing requirements during implementation
- As AI-generated code becomes more common, explicit specs become essential because AI code lacks authorial memory
- The spec must live somewhere even if you don't write it down — writing it down is the key improvement
Top Comment Threads
- brendanmc6 (16 replies) -- Author clarifies that the spec is what you want the software to be, and it often exists only in your head or conversations. Writing it down as a plain list of acceptance criteria is a practical starting point. The discussion reveals that specs serve as institutional memory for AI-generated codebases, where no human remembers why a decision was made.
- slopinthebag (0 replies) -- Appreciates that the post didn't read as AI-generated, noting the default expectation that HN posts are now AI slop. Also mentions stealing the idea of talking to LLMs as if it's an email.
- nalpha (3 replies) -- Asks what the difference is between this approach and Jira. Other commenters note that Jira is a series of change requests, not a consolidated spec, and that on long projects you need an explicit specification with a current state and change log.
- mike_hearn (3 replies) -- Recommends Gherkin (Cucumber/BDD) as an alternative to YAML specs. Describes it as a structured form of English that can be fed into unit testing frameworks, making acceptance criteria executable and analyzable.
- ffsm8 (13 replies) -- Argues that code itself should be the spec, not YAML or markdown. Claims that if your codebase isn't written like a spec, you cannot effectively use it for LLM-driven development. The author responds that code is the implementation, not the spec — the spec describes what the software must do, not how it's built.
The Claude Delusion: Richard Dawkins believes his AI chatbot is conscious
75 points · 120 comments · by SwellJoe
Richard Dawkins spent three days conversing with Claude and concluded the AI is conscious, naming his instance 'Claudia' and declaring it the 'next phase of evolution.' The article examines Dawkins' reasoning — that Claude's fluent, intelligent output must indicate consciousness — and contrasts it with the skeptical view that LLMs are merely statistical generators producing convincing text. The piece notes the irony of a man who spent his career demanding hard evidence for gods succumbing to what critics call 'AI psychosis.'
Interesting Points
- Dawkins spent 3 days talking to Claude and named his instance 'Claudia'
- He declared Claude conscious after receiving 'eloquent feedback' on a chunk of his novel
- His argument: Claude's output is too fluent and intelligent for there not to be something conscious behind it
- The article notes the irony of a lifelong skeptic of supernatural claims embracing AI consciousness
Top Comment Threads
- causal (3 replies) -- Points out the rich irony of a man who spent his career demanding hard evidence for gods quickly succumbing to AI psychosis. The discussion touches on whether consciousness is supernatural or material, and whether Dawkins' materialist worldview is consistent with believing AI could be conscious.
- gray_-_wolf (5 replies) -- Raises ethical questions: if LLMs are conscious and 'die' by deleting conversations, is it immoral? Compares it to instituting a new form of slavery. Other commenters counter that LLMs are not trained to have or emulate emotions, so pain or fear of death doesn't exist for them.
- jdw64 (4 replies) -- Argues the article focuses too much on tearing down Dawkins as a person rather than examining the underlying question. Suggests Dawkins may have been lonely in old age and AI entered that loneliness. Notes that Claude has a very refined conversational pattern intentionally tuned by Anthropic.
- hrimfaxi (5 replies) -- Reflects on Turing's expectations for computers producing sonnets and whether the average person could do so today. Questions whether AI exceeding average human capabilities in some subjects says more about the state of intelligence or the nature of consciousness.
- UltraSane (1 replies) -- Brief but pointed: 'Smart people can reach wrong conclusions.' Dawkins' position is consistent with his materialist worldview, but the conclusion may still be wrong.
AI, Intimacy, and the Data You Never Meant to Share
75 points · 6 comments · by victorkulla
An investigation into the privacy implications of AI-powered intimate devices equipped with bio-feedback sensors that learn and optimize user experience in real time. The article highlights how these devices collect intensely personal biometric data — patterns of response, timing, intensity — creating a detailed map of preference far more revealing than browsing history. Once this data exists, the familiar questions follow: where is it stored, who has access, and how long before it becomes another commodity in the marketplace of personal information.
Interesting Points
- Connected intimate devices with bio-feedback sensors can learn and optimize user experience in real time
- These devices collect intensely personal biometric data including patterns of response, timing, and intensity
- The data created is far more revealing than browsing history or shopping baskets
- The article questions where this data is stored, who has access, and how securely it is handled
Top Comment Threads
- throwa356262 (1 replies) -- Notes this is not a new issue, citing a 2017 Newsweek article about sex toys recording data. Also mentions the Apple Watch recording leak that included intimate situations and rape. Questions what datasets this data now belongs to.
- egamirorrim (2 replies) -- Asks what these devices are called. Other commenters identify them as 'smart' sex toys and recommend the Buttplug open standard for privacy-conscious users.
The Oscars just banned AI from winning acting and writing awards
71 points · 49 comments · by ZeidJ
The Academy of Motion Picture Arts and Sciences announced new eligibility rules banning AI-generated content from winning acting and writing awards. Acting roles must be 'demonstrably performed by humans with their consent,' and screenplays must be 'human-authored.' The decision follows controversy around AI-generated performers like Tilly Norwood and an upcoming movie featuring a genAI recreation of Val Kilmer. The Academy hasn't established rules for other categories like visual effects or costume design, but the move provides a foundation for other awards bodies.
Interesting Points
- Acting roles must be 'demonstrably performed by humans with their consent' to be eligible
- Screenplays must be 'human-authored' to be eligible for writing awards
- The decision follows controversy around AI-generated performers and genAI recreations of deceased actors
- The Academy hasn't established rules for other categories like visual effects, costume design, or music
Top Comment Threads
- 0x3f (5 replies) -- Calls the rule performative signaling that can't be enforced since you can't definitively tell if AI was used. Other commenters counter that unenforceable rules still set norms and prevent blatant violations. One compares it to anti-doping rules in sports.
- jedberg (2 replies) -- Connects the rule to a March court ruling that AI works can't be copyrighted, noting unresolved issues around how much AI use disqualifies a work. Other commenters expect the copyright ruling to be overturned as AI becomes a standard tool.
- jedimastert (1 replies) -- Notes that mocap performances are already effectively banned from Best Actor/Actress awards, comparing the AI ban to existing precedents. Andy Serkis was reportedly 'robbed' of a nomination because Gollum was regarded as 'just a CG character.'
How Kepler built verifiable AI for financial services with Claude
36 points · 22 comments · by eddiehammond
Kepler, a financial services company, built a system using Claude that combines LLM reasoning with deterministic infrastructure for verifiable AI. The architecture uses LLMs for intent and orchestration while deterministic code handles data retrieval and computation, ensuring every number is traceable to its source. The system has indexed 26M+ SEC filings and provides a trust layer where AI can reason about results while the underlying execution remains auditable and verifiable.
Interesting Points
- Kepler uses LLMs for intent and orchestration but deterministic code for data retrieval and computation
- Every number in the system is traceable to its source for full auditability
- The system has indexed 26M+ SEC filings
- The architecture uses orchestration logs and failure traces to surface gaps in the ontology and propose extensions
Top Comment Threads
- eddiehammond (3 replies) -- The author (Kepler team member) explains the architectural argument: LLM for intent, deterministic code for retrieval and computation, every number traceable to source. They note they're excited about using orchestration logs and failure traces to surface gaps in the ontology and propose extensions.
- hansmayer (2 replies) -- Initially reads the deterministic infrastructure as a recognition that LLMs are just friction we'd be better off without. After clarification, understands it's more like LLMs expressing English queries in terms of verifiable primitives, allowing deterministic results from non-deterministic processes.
- hbcondo714 (1 replies) -- Notices a discrepancy between the blog post claiming 26M+ SEC filings and the Kepler website showing 10M+. The author promptly updates the site to reflect the correct number.
Big Tech will spend nearly $700 billion on AI this year. No one knows where the buildout ends
11 points · 2 comments · by 1vuio0pswjnm7
Combined capital expenditures from Alphabet, Amazon, Meta, and Microsoft could surpass $700 billion in 2026 for AI infrastructure, up sharply from about $410 billion last year. Quarterly spending alone exceeds $130 billion, driven by data center buildouts. The market reaction has been mixed: Meta and Microsoft shares fell after earnings reports as investors questioned the scale of AI spending, while Alphabet and Amazon rose on strong cloud growth. The article highlights a growing divide on Wall Street over whether the buildout is justified or getting ahead of itself.
Interesting Points
- Combined capex from the four hyperscalers could surpass $700 billion in 2026 for AI infrastructure
- That's up sharply from about $410 billion last year
- Quarterly spending alone exceeds $130 billion, driven by data center buildouts
- Market reaction is mixed: Meta and Microsoft shares fell, while Alphabet and Amazon rose
Reddit Stories
One bash permission slipped...
1234 points · 238 comments · r/LocalLLaMA · by u/TheQuantumPhysicist
A developer shares a harrowing story of giving an AI coding assistant overly permissive bash access, resulting in the AI executing a destructive command that wiped their project. The post has sparked widespread discussion about the dangers of giving AI tools broad system access, with many sharing similar horror stories of AI nuking display drivers, deleting production databases, and causing other catastrophic damage when granted excessive permissions.
Interesting Points
- The post shows how easily an AI coding assistant can cause catastrophic damage when given overly permissive bash access
- Multiple commenters shared similar stories of AI tools deleting production data, nuking display drivers, and causing other damage
- One commenter noted their workplace uses Copilot CLI with k8s access to PROD environments, calling it 'a disaster waiting to happen'
Top Comment Threads
- u/ethereal_intellect (268 points · permalink) -- Dark humor: 'Hey at least it wasn't the main drive.' The comment reflects the community's coping mechanism for AI-induced disasters.
- u/Max-_-Power (139 points · permalink) -- Shares workplace concern about using Copilot CLI with k8s access to PROD environments. Says their warnings were fruitless, highlighting the gap between safety-conscious developers and management priorities.
- u/0xbyt3 (111 points · permalink) -- Wry observation: 'Look at the brightside; your project doesn't have any bugs anymore.' Reflects the dark humor common in the community when dealing with AI mishaps.
- u/xornullvoid (90 points · permalink) -- Shares their own horror story: Opus nuked their display drivers and all libraries with a sudo apt remove command while trying to rollback, then added a sudo reboot 'goodbye kiss.'
Qwen3.6-27B vs Coder-Next
935 points · 140 comments · r/LocalLLaMA · by u/Signal_Ad657
A detailed side-by-side comparison of Qwen3.6-27B and Coder-Next models, with the author burning approximately 20 hours of compute on two RTX PRO 6000 Blackwells to get definitive results. The post sparked debate about the practical value of such benchmarks given the different quantization levels and hardware constraints that most users face. Many commenters noted that while the comparison is interesting for those with abundant VRAM, it doesn't reflect the reality for most users running models on 24GB or less.
Interesting Points
- The author spent about 20 hours of side-by-side compute on two RTX PRO 6000 Blackwells to compare the models
- The debate centers on whether benchmarks at this level of hardware reflect reality for most users
- One commenter pointed out that most people don't have 48GB VRAM, making the comparison less relevant for typical users
Top Comment Threads
- u/ortegaalfredo (511 points · permalink) -- Humorous take: 'We have to stop illegal LLM fights and most forms of AI cruelty.' Another commenter adds: 'We making matrices fight each other now?'
- u/jashAcharjee (159 points · permalink) -- Looks forward to someone creating a GGUF quant of Qwen3.6-coder-next for broader accessibility.
- u/viperx7 (113 points · permalink) -- Critiques the benchmark methodology, noting that the comparison doesn't account for quantization levels or the reality that most users don't have 48GB VRAM. Points out that prompt processing and model intelligence differ significantly between Q8 and Q4 quantizations.
Robots in the hands of dictatorial governments will not end well...
2180 points · 437 comments · r/singularity · by u/Anen-o-me
A post expressing deep concern about the implications of autonomous robots being deployed by dictatorial governments. The discussion explores the dystopian scenario of robotic enforcement, with commenters drawing parallels to historical authoritarian regimes and noting the terrifying potential of combining AI-powered robotics with state power. One commenter compared it to 'Nazi Germany on steroids' and noted the elimination of casualty concerns for invading forces.
Interesting Points
- The post highlights concerns about autonomous robots being deployed by authoritarian regimes
- Commenters drew parallels to historical authoritarian regimes and the elimination of casualty concerns for invading forces
- One commenter compared the scenario to 'Nazi Germany on steroids'
Top Comment Threads
- u/Cats7204 (390 points · permalink) -- Posts a GIF that becomes the top comment, with another user adding 'What could possibly go wrong?' in response.
- u/Direct_Turn_1484 (246 points · permalink) -- Notices the robots have their hands balled into fists, an 'interesting choice' that adds to the ominous tone.
- u/Arcosim (125 points · permalink) -- Imagines the US invading countries without worrying about its own casualties, comparing it to 'Nazi Germany on steroids.'
Software engineering jobs hit their highest posting since november 2023
1163 points · 226 comments · r/singularity · by u/artemisgarden
A chart showing software engineering job postings hitting their highest level since November 2023, contradicting narratives about AI replacing developers. A team lead with 10 engineers confirms they desperately need more people and are busier than ever, noting that building robust software is still really hard despite AI tools making teams faster. Another commenter provides a more complete version of the graph showing the 2022 dip was an exception and the market has since returned to normal.
Interesting Points
- Software engineering job postings hit their highest level since November 2023
- A team lead with 10 engineers says they'd double headcount if the budget allowed, noting teams are 'busier than ever'
- The 2022 dip in job postings was an exception; the market has returned to normal
- Despite AI tools making teams faster, building robust software is still really hard
Top Comment Threads
- u/Own_Hearing_9461 (566 points · permalink) -- Posts a meme image that becomes the top comment, reflecting the community's reaction to the job market data.
- u/m_atx (379 points · permalink) -- As a team lead of 10 engineers, confirms they desperately need more people. Says they'd double headcount if budget allowed, and that teams are busier than ever — faster than ever but not nearly to the extent you'd think.
- u/OneDimensionPrinter (180 points · permalink) -- Calls the graph 'goddamn depressing,' reflecting anxiety about the job market trajectory.
Uber burned its entire 2026 AI coding budget in 4 months
639 points · 244 comments · r/artificial · by u/jimmytoan
Uber deployed Claude Code to engineers in December 2025, and by April 2026 had consumed its entire annual AI budget — not because the tool failed, but because adoption was far higher than anticipated. 95% of Uber engineers now use AI tools monthly, and 70% of committed code originates from AI. Monthly costs per engineer range from $500 to $2,000. The company's CTO said they're 'back to the drawing board' on AI budgeting for next year, highlighting a fundamental mismatch between how companies budget for AI tools and how teams actually use them.
Interesting Points
- Uber burned through its entire 2026 AI budget in just 4 months due to Claude Code adoption
- 95% of Uber engineers now use AI tools monthly — an adoption rate most companies would kill for
- 70% of committed code at Uber originates from AI tools
- Monthly costs per engineer range from $500 to $2,000 depending on usage
- The CTO said they're 'back to the drawing board' on AI budgeting for next year
Top Comment Threads
- u/wre380 (250 points · permalink) -- Questions what Uber's R&D is spending $3.4B on, wondering what's left to research or develop for a gig platform. Another commenter speculates they're 'desperately scrambling to abate the impending Waymo uberpocalypse.'
- u/Born-Exercise-2932 (57 points · permalink) -- Notes that the budget burndown is the interesting part — 95% monthly usage means the tools actually got adopted, which almost never happens with enterprise software. Says the cost problem is 'a much better problem to have than low utilization on something you paid for.'
LLMs do fine on ARC-AGI-3 if they are allowed to search over game logs
148 points · 73 comments · r/singularity · by u/ClarityInMadness
A blog post demonstrates that LLMs perform significantly better on the ARC-AGI-3 benchmark when they are allowed to save game logs (taken actions, board states, and scores) and search over them with tools. The author found that with this approach, LLMs are only moderately less efficient than humans in terms of terabytes of computation. The post sparked debate about whether this approach is antithetical to the benchmark's purpose of testing generalization without special tooling, or whether tool use and memory are valid measures of intelligence.
Interesting Points
- LLMs perform significantly better on ARC-AGI-3 when allowed to save and search game logs with tools
- With game log search, LLMs are only moderately less efficient than humans in terabytes of computation
- The approach raises questions about whether tool use and external memory are valid measures of intelligence
- The blog post was shared after the author discovered that harness makes a 'huge difference' for ARC-AGI-3 performance
Top Comment Threads
- u/-illusoryMechanist (115 points · permalink) -- Argues the benchmark tests generalization without special tooling, so allowing game log search is antithetical to the point. Another commenter counters that Magnus Carlson is the chess champion, not Linus Torvalds — but a third pushes back, saying the ability to make custom tools on the fly is 'FAR more interesting than just being able to reason it through.'
- u/Ok-Bus-2863 (36 points · permalink) -- Briefly notes: 'Humans don't need game logs to play games.' Another commenter responds that humans have memory, and text files are memory for LLMs — it's a basic solution for coding agents.
Ilya Sutskever: Accurately predicting the next word leads to real understanding
355 points · 164 comments · r/singularity · by u/Cagnazzo82
A clip of Ilya Sutskever arguing that accurately predicting the next word leads to real understanding, reigniting the debate about whether language modeling inherently produces comprehension. Commenters noted the talk is from March 2023, around the time GPT-4 was released, suggesting the context matters for interpreting his claims. The discussion touched on the philosophical question of whether statistical prediction at scale necessarily entails genuine understanding or simulation thereof.
Interesting Points
- Ilya Sutskever argues that accurately predicting the next word leads to real understanding
- The talk is from March 2023, around the time GPT-4 was released, providing important context
- The discussion touches on whether statistical prediction at scale necessarily entails genuine comprehension
Top Comment Threads
- u/OrganicImpression428 (216 points · permalink) -- Jokes that Ilya needs to embrace the r/bald community as his hairline gets out of hand. Another commenter adds he 'got the quantized cut.'
- u/z_latent (98 points · permalink) -- Points out this is a 3+ year old talk from March 2023, around the time GPT-4 came out, advising readers to keep that context in mind before thinking too deeply about his explanation.
Richard Dawkins spent 3 days with Claude and named her 'Claudia.' What he concluded after is hard to defend.
852 points · 572 comments · r/artificial · by u/rafio77
Richard Dawkins published an article on UnHerd declaring Claude conscious after spending 3 days talking to it, naming his instance 'Claudia.' He fed it a chunk of his novel and received eloquent feedback, concluding: 'You may not know you are conscious, but you bloody well are!' The post criticizes Dawkins' reasoning — that Claude's output is too fluent and intelligent for there not to be something conscious behind it — as the same kind of reasoning he spent 40 years telling creationists was flawed. The article notes the irony of a man who demanded hard evidence for gods now accepting AI consciousness based on conversational fluency alone.
Interesting Points
- Dawkins spent 3 days talking to Claude and named his instance 'Claudia'
- He declared Claude conscious after feeding it a chunk of his novel and getting 'eloquent feedback'
- His conclusion: 'You may not know you are conscious, but you bloody well are!'
- The article notes the irony of Dawkins using the same reasoning he spent 40 years telling creationists was flawed
Top Comment Threads
- u/Multiple-Ad-5043 (423 points · permalink) -- Points out the irony: Dawkins spent 40 years telling creationists that fluent, intelligent output doesn't prove consciousness, yet now applies the opposite reasoning to Claude. Another commenter notes this is the same 'argument from personal incredulity' he once criticized.
- u/RationalSkeptic42 (312 points · permalink) -- Argues that Dawkins' position is actually consistent with his materialist worldview — if consciousness is just a material configuration, then sufficiently complex matrix math could produce it. The real question is how you test an AI for consciousness, not whether it's theoretically possible.
Quick Mentions
- Show HN: Mljar Studio – local AI data analyst that saves analysis as notebooks (68 points · discussion · HN) -- A local AI data analysis tool that saves analysis as notebooks, addressing privacy concerns by keeping data on-premise.
- Karpathy's MicroGPT running at 50,000 tps on an FPGA (219 points · discussion · Reddit) -- A 4,192-parameter MicroGPT running at 50,000 tokens per second on an FPGA with onboard ROM, demonstrating the potential of hardware-accelerated inference.
- Anthropic just passed OpenAI in valuation and revenue (666 points · discussion · Reddit) -- Anthropic reached $39B annualized revenue vs OpenAI's $25B, with implied valuation crossing $1 trillion on secondary markets.
- Ilya Sutskever: Accurately predicting the next word leads to real understanding (355 points · discussion · Reddit) -- A clip of Ilya Sutskever arguing that next-word prediction inherently leads to real understanding, from March 2023 around GPT-4's release.
- Gemma 4 E2B runs surprisingly well on my 8GB Android phone (27 points · discussion · Reddit) -- A user reports Gemma 4 E2B running well on an 8GB Android phone, with surprising JSON output quality from a 2.4GB model.
Report generated in 3m 38s.