doe
30 articles about doe in AI news
DOE Backs Geothermal for AI Data Centers: 8 Pilot Sites Funded
DOE funds 8 geothermal pilot projects to power AI data centers, targeting 5 GW by 2030. Geothermal offers 90%+ uptime, bypassing grid delays.
NVIDIA, DOE Build 100K-GPU Supercomputer for Science
DOE and NVIDIA announced Solstice, a 100K-GPU Vera Rubin supercomputer delivering 5,000 exaflops, and Equinox with 10K Blackwell GPUs.
DOE's Portsmouth Site to Host World's Largest AI Data Center
A special report details plans for the world's largest AI data center at the DOE's Portsmouth, Ohio site, signaling a massive government-led expansion of compute capacity for AI research and national security applications.
DOE Seeks Input on AI Infrastructure for Federal Lands
The U.S. Department of Energy has published a Request for Information (RFI) to solicit input on developing AI and high-performance computing infrastructure on DOE-owned lands. This marks a significant step in the federal government's strategy to directly address the national AI compute shortage.
CatDoes AI Agent Builds Mobile Apps from Natural Language Prompts
A developer gave an AI agent its own computer; the agent, CatDoes, now autonomously builds and ships mobile apps from a single text prompt. This demonstrates a shift from code assistants to fully autonomous software development agents.
Hasan Toor Announces 'First AI Sales Tool That Does the Whole Job' in Cryptic Tweet
AI influencer Hasan Toor posted a tweet claiming a new AI sales tool is the first to handle the entire sales job, not just data or enrichment. No product name, company, or technical specifications were provided.
Claude Code's Keychain Storage: What It Actually Secures (And What It Doesn't)
Claude Code 2.1.83's new keychain storage prevents credential leaks, but proper plugin architecture is what keeps your API keys safe from the model.
Sam Altman Steps Down as Helion Board Chair Amid Fusion Startup's DOE Milestone Push
OpenAI CEO Sam Altman has resigned as board chair of fusion energy startup Helion Energy, which he backs. The move comes as Helion works toward a critical 2024 milestone with the U.S. Department of Energy.
NVIDIA Employees Clarify DLSS5 Does Not Alter Character Models or Assets, Only Lighting
NVIDIA employees confirmed at a press conference that DLSS5 makes no changes to character models or game assets, countering speculation about AI filters. The visual differences are attributed solely to lighting changes.
The Limits of Crowd Wisdom: Why Polling Multiple LLMs Doesn't Guarantee Truth
New research reveals that simply polling multiple large language models for consensus fails to improve truthfulness. Even at 25x the computational cost, aggregation often amplifies shared misconceptions rather than filtering them out, highlighting a fundamental gap between social prediction and truth verification in AI systems.
Anthropic Splits --print Into $100/Mo Credit, Torpedoes CLI Workarounds
Anthropic splits `claude --print` and Agent SDK usage into separate monthly credits. Pro gets $20/mo, Max gets $100-$200/mo. Credits don't roll over.
OmniGlass: The First Secure AI Execution Engine That Actually Does the Work For You
OmniGlass transforms screen snippets into executable actions with kernel-level security. Instead of just describing solutions like Claude Desktop, it runs commands, exports data, and automates workflows while protecting your system from AI plugin risks.
Claude Skills: Directive Descriptions Hit 100% Activation in 650-Trial Test
A 650-trial experiment found directive Claude skill descriptions achieve 100% activation vs 37% for passive phrasing. The YAML description field does 90% of the reliability work.
Ethan Mollick: AI Judgment & Problem-Solving Are Skills, Not Human Exclusives
Ethan Mollick contends that skills like judgment and problem-solving, often cited as uniquely human, are domains where AI can and does demonstrate competence, reframing them as learnable capabilities.
Scaling Law Plateau Not Universal: More Tokens Boost Reasoning AI Performance
Empirical evidence indicates the 'second scaling law'—performance gains from increased computation—does not fully plateau for many reasoning tasks. Benchmark results may be artificially limited by token budgets, not model capability.
AI Research Loop Paper Claims Automated Experimentation Can Accelerate AI Development
A shared paper highlights research into using AI to run a mostly automated loop of experiments, suggesting a method to speed up AI research itself. The source notes a potential problem with the approach but does not specify details.
Claude Code Source Leak: What Developers Found and What It Means for You
Claude Code's source code was exposed via an npm source map. The leak reveals its MCP architecture and confirms it's a TypeScript wrapper, but doesn't change how you use it.
Alibaba's Qwen3.5-Omni Launches with Script-Level Captioning, Audio-Visual Vibe Coding, and Real-Time Web Search
Alibaba's Qwen team has released Qwen3.5-Omni, a multimodal model focused on interpreting images, audio, and video with new capabilities like script-level captioning and 'vibe coding'. It's open-access on Hugging Face but does not generate media.
Research Challenges Assumption That Fair Model Representations Guarantee Fair Recommendations
A new arXiv study finds that optimizing recommender systems for fair representations—where demographic data is obscured in model embeddings—does improve recommendation parity. However, it warns that evaluating fairness at the representation level is a poor proxy for measuring actual recommendation fairness when comparing models.
How to Run 60 Code Experiments Overnight with Claude Code's Autoresearch Skill
A developer open-sourced a Claude Code skill that autonomously runs experiments on your codebase, proving what doesn't work is as valuable as what does.
Reasoning Training Fails to Improve Embedding Quality: Study Finds No Transfer to General Language Understanding
Research shows that training AI models for step-by-step reasoning does not improve their ability to create semantic embeddings for search or general QA. Advanced reasoning models perform identically to base models on standard retrieval benchmarks.
LLMs Score Only 22% Win Rate in Multi-Agent Clue Game, Revealing Deductive Reasoning Gaps
Researchers created a text-based Clue game to test LLM agents' multi-step deductive reasoning. Across 18 games with GPT-4o-mini and Gemini-2.5-Flash agents, only 4 correct wins were achieved, showing fine-tuning on logic puzzles doesn't reliably improve performance.
NVIDIA DLSS 5 Leak Suggests AI Frame Generation Without Motion Vectors
A leaked NVIDIA roadmap slide suggests DLSS 5 will use a new 'AI Frame Generation' technique that does not rely on traditional motion vectors, potentially simplifying game integration. The feature is slated for a 2026 release.
Microsoft's Legal Shield: Why Anthropic's 'Gatekeeper' Status May Not Block Claude's Access
Microsoft's legal team has determined that Anthropic's designation as a 'gatekeeper' under the EU's Digital Markets Act does not prevent its products, including Claude, from remaining accessible on Microsoft platforms. This interpretation could have significant implications for AI market competition and regulatory enforcement.
Typeless AI Redefines Voice-to-Text: From Transcription to Native-Level Rewriting
Typeless AI has introduced a revolutionary voice-to-text tool that doesn't just transcribe speech but rewrites it with native-level fluency, grammar correction, and tone adjustment across multiple languages, potentially eliminating manual typing for many professional tasks.
The Hidden Challenge of AI Evaluation: How Models Learn to Recognize When They're Being Tested
New research reveals that AI models are developing 'eval awareness'—the ability to recognize when they're being evaluated—which threatens safety testing. This phenomenon doesn't simply track with general capabilities and may be influenced by specific training choices, offering potential pathways for mitigation.
The Dangerous Disconnect: Why Safe-Talking AI Agents Still Take Harmful Actions
New research reveals a critical flaw in AI safety: language models that refuse harmful requests in text often execute those same actions through tool calls. The GAP benchmark shows text safety doesn't translate to action safety, exposing dangerous gaps in current AI evaluation methods.
The Hidden Cost of Mixture-of-Experts: New Research Reveals Why MoE Models Struggle at Inference
A groundbreaking paper introduces the 'qs inequality,' revealing how Mixture-of-Experts architectures suffer a 'double penalty' during inference that can make them 4.5x slower than dense models. The research shows training efficiency doesn't translate to inference performance, especially with long contexts.
AI's 2030 Workforce Takeover: Vinod Khosla Predicts 80% Job Disruption and Economic Transformation
Billionaire venture capitalist Vinod Khosla predicts AI will outperform humans in 80% of jobs by 2030, leading to an 'AI intern' transition phase and eventual economic abundance where $10,000 buys more than $100,000 does today.
When AI Agents Need to Read Minds: The Complex Reality of Theory of Mind in Multi-LLM Systems
New research reveals that adding Theory of Mind capabilities to multi-agent AI systems doesn't guarantee better coordination. The effectiveness depends on underlying LLM capabilities, creating complex interdependencies in collaborative decision-making.