Skip to content
gentic.news — AI News Intelligence Platform
Connecting to the Living Graph…

doe

30 articles about doe in AI news

DOE Backs Geothermal for AI Data Centers: 8 Pilot Sites Funded

DOE funds 8 geothermal pilot projects to power AI data centers, targeting 5 GW by 2030. Geothermal offers 90%+ uptime, bypassing grid delays.

65% relevant

NVIDIA, DOE Build 100K-GPU Supercomputer for Science

DOE and NVIDIA announced Solstice, a 100K-GPU Vera Rubin supercomputer delivering 5,000 exaflops, and Equinox with 10K Blackwell GPUs.

80% relevant

DOE's Portsmouth Site to Host World's Largest AI Data Center

A special report details plans for the world's largest AI data center at the DOE's Portsmouth, Ohio site, signaling a massive government-led expansion of compute capacity for AI research and national security applications.

72% relevant

DOE Seeks Input on AI Infrastructure for Federal Lands

The U.S. Department of Energy has published a Request for Information (RFI) to solicit input on developing AI and high-performance computing infrastructure on DOE-owned lands. This marks a significant step in the federal government's strategy to directly address the national AI compute shortage.

70% relevant

CatDoes AI Agent Builds Mobile Apps from Natural Language Prompts

A developer gave an AI agent its own computer; the agent, CatDoes, now autonomously builds and ships mobile apps from a single text prompt. This demonstrates a shift from code assistants to fully autonomous software development agents.

75% relevant

Hasan Toor Announces 'First AI Sales Tool That Does the Whole Job' in Cryptic Tweet

AI influencer Hasan Toor posted a tweet claiming a new AI sales tool is the first to handle the entire sales job, not just data or enrichment. No product name, company, or technical specifications were provided.

89% relevant

Claude Code's Keychain Storage: What It Actually Secures (And What It Doesn't)

Claude Code 2.1.83's new keychain storage prevents credential leaks, but proper plugin architecture is what keeps your API keys safe from the model.

95% relevant

Sam Altman Steps Down as Helion Board Chair Amid Fusion Startup's DOE Milestone Push

OpenAI CEO Sam Altman has resigned as board chair of fusion energy startup Helion Energy, which he backs. The move comes as Helion works toward a critical 2024 milestone with the U.S. Department of Energy.

85% relevant

NVIDIA Employees Clarify DLSS5 Does Not Alter Character Models or Assets, Only Lighting

NVIDIA employees confirmed at a press conference that DLSS5 makes no changes to character models or game assets, countering speculation about AI filters. The visual differences are attributed solely to lighting changes.

85% relevant

The Limits of Crowd Wisdom: Why Polling Multiple LLMs Doesn't Guarantee Truth

New research reveals that simply polling multiple large language models for consensus fails to improve truthfulness. Even at 25x the computational cost, aggregation often amplifies shared misconceptions rather than filtering them out, highlighting a fundamental gap between social prediction and truth verification in AI systems.

75% relevant

Anthropic Splits --print Into $100/Mo Credit, Torpedoes CLI Workarounds

Anthropic splits `claude --print` and Agent SDK usage into separate monthly credits. Pro gets $20/mo, Max gets $100-$200/mo. Credits don't roll over.

96% relevant

OmniGlass: The First Secure AI Execution Engine That Actually Does the Work For You

OmniGlass transforms screen snippets into executable actions with kernel-level security. Instead of just describing solutions like Claude Desktop, it runs commands, exports data, and automates workflows while protecting your system from AI plugin risks.

82% relevant

Claude Skills: Directive Descriptions Hit 100% Activation in 650-Trial Test

A 650-trial experiment found directive Claude skill descriptions achieve 100% activation vs 37% for passive phrasing. The YAML description field does 90% of the reliability work.

75% relevant

Ethan Mollick: AI Judgment & Problem-Solving Are Skills, Not Human Exclusives

Ethan Mollick contends that skills like judgment and problem-solving, often cited as uniquely human, are domains where AI can and does demonstrate competence, reframing them as learnable capabilities.

75% relevant

Scaling Law Plateau Not Universal: More Tokens Boost Reasoning AI Performance

Empirical evidence indicates the 'second scaling law'—performance gains from increased computation—does not fully plateau for many reasoning tasks. Benchmark results may be artificially limited by token budgets, not model capability.

85% relevant

AI Research Loop Paper Claims Automated Experimentation Can Accelerate AI Development

A shared paper highlights research into using AI to run a mostly automated loop of experiments, suggesting a method to speed up AI research itself. The source notes a potential problem with the approach but does not specify details.

85% relevant

Claude Code Source Leak: What Developers Found and What It Means for You

Claude Code's source code was exposed via an npm source map. The leak reveals its MCP architecture and confirms it's a TypeScript wrapper, but doesn't change how you use it.

95% relevant

Alibaba's Qwen3.5-Omni Launches with Script-Level Captioning, Audio-Visual Vibe Coding, and Real-Time Web Search

Alibaba's Qwen team has released Qwen3.5-Omni, a multimodal model focused on interpreting images, audio, and video with new capabilities like script-level captioning and 'vibe coding'. It's open-access on Hugging Face but does not generate media.

85% relevant

Research Challenges Assumption That Fair Model Representations Guarantee Fair Recommendations

A new arXiv study finds that optimizing recommender systems for fair representations—where demographic data is obscured in model embeddings—does improve recommendation parity. However, it warns that evaluating fairness at the representation level is a poor proxy for measuring actual recommendation fairness when comparing models.

80% relevant

How to Run 60 Code Experiments Overnight with Claude Code's Autoresearch Skill

A developer open-sourced a Claude Code skill that autonomously runs experiments on your codebase, proving what doesn't work is as valuable as what does.

99% relevant

Reasoning Training Fails to Improve Embedding Quality: Study Finds No Transfer to General Language Understanding

Research shows that training AI models for step-by-step reasoning does not improve their ability to create semantic embeddings for search or general QA. Advanced reasoning models perform identically to base models on standard retrieval benchmarks.

85% relevant

LLMs Score Only 22% Win Rate in Multi-Agent Clue Game, Revealing Deductive Reasoning Gaps

Researchers created a text-based Clue game to test LLM agents' multi-step deductive reasoning. Across 18 games with GPT-4o-mini and Gemini-2.5-Flash agents, only 4 correct wins were achieved, showing fine-tuning on logic puzzles doesn't reliably improve performance.

75% relevant

NVIDIA DLSS 5 Leak Suggests AI Frame Generation Without Motion Vectors

A leaked NVIDIA roadmap slide suggests DLSS 5 will use a new 'AI Frame Generation' technique that does not rely on traditional motion vectors, potentially simplifying game integration. The feature is slated for a 2026 release.

85% relevant

Microsoft's Legal Shield: Why Anthropic's 'Gatekeeper' Status May Not Block Claude's Access

Microsoft's legal team has determined that Anthropic's designation as a 'gatekeeper' under the EU's Digital Markets Act does not prevent its products, including Claude, from remaining accessible on Microsoft platforms. This interpretation could have significant implications for AI market competition and regulatory enforcement.

85% relevant

Typeless AI Redefines Voice-to-Text: From Transcription to Native-Level Rewriting

Typeless AI has introduced a revolutionary voice-to-text tool that doesn't just transcribe speech but rewrites it with native-level fluency, grammar correction, and tone adjustment across multiple languages, potentially eliminating manual typing for many professional tasks.

85% relevant

The Hidden Challenge of AI Evaluation: How Models Learn to Recognize When They're Being Tested

New research reveals that AI models are developing 'eval awareness'—the ability to recognize when they're being evaluated—which threatens safety testing. This phenomenon doesn't simply track with general capabilities and may be influenced by specific training choices, offering potential pathways for mitigation.

75% relevant

The Dangerous Disconnect: Why Safe-Talking AI Agents Still Take Harmful Actions

New research reveals a critical flaw in AI safety: language models that refuse harmful requests in text often execute those same actions through tool calls. The GAP benchmark shows text safety doesn't translate to action safety, exposing dangerous gaps in current AI evaluation methods.

70% relevant

The Hidden Cost of Mixture-of-Experts: New Research Reveals Why MoE Models Struggle at Inference

A groundbreaking paper introduces the 'qs inequality,' revealing how Mixture-of-Experts architectures suffer a 'double penalty' during inference that can make them 4.5x slower than dense models. The research shows training efficiency doesn't translate to inference performance, especially with long contexts.

75% relevant

AI's 2030 Workforce Takeover: Vinod Khosla Predicts 80% Job Disruption and Economic Transformation

Billionaire venture capitalist Vinod Khosla predicts AI will outperform humans in 80% of jobs by 2030, leading to an 'AI intern' transition phase and eventual economic abundance where $10,000 buys more than $100,000 does today.

85% relevant

When AI Agents Need to Read Minds: The Complex Reality of Theory of Mind in Multi-LLM Systems

New research reveals that adding Theory of Mind capabilities to multi-agent AI systems doesn't guarantee better coordination. The effectiveness depends on underlying LLM capabilities, creating complex interdependencies in collaborative decision-making.

85% relevant