amd
30 articles about amd in AI news
AMD AI Director Reports Claude Code Quality Decline, Cites 234k Tool Calls
An AMD AI executive presented data from over 6,800 sessions showing Claude Code's performance has declined since early March, with rising instances of shallow reasoning and incomplete tasks. This raises significant trust issues for engineers using the model in complex development workflows.
MLPerf 6.0: NVIDIA Sweeps New Benchmarks, AMD MI355X Within 30% on Select Tests
MLPerf 6.0 results show NVIDIA winning every new benchmark, with its GB300 NVL72 system achieving nearly 3x more throughput than six months ago. AMD's MI355X showed progress, coming within 10-30% on select single-node tests but skipping most new benchmarks.
Developer Ranks NPU Model Compilation Ease: Apple 1st, AMD Last
Developer @mweinbach ranked the ease of using AI coding agents to compile ML models for NPUs. Apple's ecosystem was rated easiest, while AMD's tooling was ranked most difficult.
Meta's $100B AMD Gamble: The AI Chip War Enters Its Most Strategic Phase
Meta has secured a landmark deal to purchase up to $100 billion worth of AMD AI chips, receiving a massive stock warrant in return. This unprecedented agreement signals Meta's aggressive push to diversify its AI infrastructure beyond Nvidia while pursuing ambitious 'personal superintelligence' goals.
Meta's $100 Billion AMD Bet: The AI Infrastructure Arms Race Reaches New Heights
Meta has reportedly signed a staggering $100 billion agreement with AMD to secure 6GW of data center capacity, signaling an unprecedented commitment to AI infrastructure. The timing—just before NVIDIA's quarterly results—highlights intensifying competition for computing resources essential for next-generation AI models.
NVIDIA's DreamDojo: Teaching Robots to 'Dream' in Pixels with 44,000 Hours of Human Experience
NVIDIA has open-sourced DreamDojo, a revolutionary robot world model trained on 44,711 hours of real-world human video. Instead of relying on physics engines, it predicts action outcomes directly in pixel space, potentially accelerating robotics development by orders of magnitude.
DeepSeek V4 Launch Signals China's Strategic Shift in AI Chip Independence
DeepSeek's upcoming V4 multimodal model prioritizes domestic chip partners Huawei and Cambricon over NVIDIA and AMD, marking a significant move toward Chinese AI self-sufficiency amid ongoing U.S. export restrictions.
Anthropic Considers Custom AI Chips, Following Google & OpenAI
Anthropic is reportedly considering developing custom AI chips, a strategic move to gain control over its compute infrastructure and reduce costs. This follows similar initiatives by Google, Amazon, and OpenAI.
Intel & Google Announce Multiyear AI & Cloud Infrastructure Partnership
Intel and Google have announced a multiyear strategic collaboration to advance AI and cloud infrastructure, focusing on optimizing Google Cloud for Intel's Xeon processors, Gaudi AI accelerators, and future chips.
Intel, SambaNova Blueprint Pairs GPUs for AI Prefill, RDUs for Decoding
Intel and SambaNova Systems have outlined a new inference architecture for agentic AI workloads. It splits tasks between GPUs for 'prefill' and SambaNova's Reconfigurable Dataflow Units (RDUs) for high-throughput token generation.
Broadcom to Manufacture Google TPU Chips in Foundry Partnership
Google has licensed its Tensor Processing Unit (TPU) intellectual property to Broadcom for chip fabrication. This allows Google to earn from its IP while Broadcom manages the complex hardware build and networking integration.
Dell XPS 14 with Core Ultra X7 Outlasts Snapdragon X Elite in Battery Test
A new battery test shows the Dell XPS 14 with Intel Core Ultra X7 lasts 11.7 hours, beating the 11.3 hours of Microsoft's ARM-based Surface Laptop 15 with Snapdragon X Elite. This is a significant win for Intel's latest chip in the critical mobile performance metric.
Qualcomm X2 Elite Matches Apple M5 in Efficiency Test
In a mixed-use laptop test simulating office work, Qualcomm's Snapdragon X2 Elite system-on-chip matched the power efficiency of Apple's latest M5 chip. This marks a significant milestone for Windows on Arm in its competition with Apple Silicon.
Terafab's 1GW AI Compute Goal Requires Massive Fab Capacity
Analysis of Terafab's stated goals shows that achieving 1GW of AI compute would require approximately 190,000 wafer starts per month across logic and memory. This underscores the unprecedented scale of semiconductor manufacturing needed for future AI infrastructure.
Microsoft's BitNet Enables 100B-Parameter LLMs on CPU, Cuts Energy 82%
Microsoft Research's BitNet project demonstrates 1-bit LLMs with 100B parameters that run efficiently on CPUs, using 82% less energy while maintaining performance, challenging the need for GPUs in local deployment.
Intel Joins SpaceX, xAI, Tesla in 'Terafab' Chip Project
Intel announced it is joining the 'Terafab' project alongside SpaceX, xAI, and Tesla. The collaboration aims to refactor silicon fab technology, likely to support the massive compute demands of AI and aerospace.
FDA-Designated AI 'Vox' Detects Heart Failure from 5-Second Voice Clip
An AI tool named Vox can detect signs of worsening heart failure from a 5-second patient voice clip. It's trained on >3M voice samples and backed by five clinical trials, targeting a condition affecting 64M people globally.
McKinsey: AI Infrastructure Value Creation Outpaces Business Capture
McKinsey's latest analysis indicates the pace of value creation from AI infrastructure is exceeding the rate at which most businesses are capturing it, highlighting a growing implementation deficit.
US Data Center Power Demand Hits 15 GW, Grid Constraints Emerge
US data center power demand reached 15 gigawatts in 2023, up from 11 GW in 2022. This rapid growth highlights a widening bottleneck: compute infrastructure is scaling faster than power delivery systems can support.
OpenAI, Anthropic Forecast $121B Compute Burn, Revealing AI's True Cost
Internal forecasts from OpenAI and Anthropic reveal the core challenge of modern AI has shifted from selling the technology to financing the immense compute required for training and inference, with OpenAI projecting $121B in compute spending for 2028.
Neuromorphic Computing Patents Surge 401% in 2025, Hits 596 by 2026
Patent filings for neuromorphic computing—hardware that mimics the brain's architecture—surged 401% in 2025, reaching 596 by early 2026. This indicates the technology is transitioning from lab prototypes to commercial products.
Chamath Palihapitiya: AI's Biggest Profits Won't Go to Model Makers
VC Chamath Palihapitiya posits that the greatest financial winners in AI will be application builders with unique distribution, not the foundational model creators, drawing a parallel to refrigeration and Coca-Cola.
X Post Reveals Audible Quality Differences in GPU vs. NPU AI Inference
A developer demonstrated audible quality differences in AI text-to-speech output when run on GPU, CPU, and NPU hardware, highlighting a key efficiency vs. fidelity trade-off for on-device AI.
PicoClaw: $10 RISC-V AI Agent Challenges OpenClaw's $599 Mac Mini Requirement
Developers have launched PicoClaw, a $10 RISC-V alternative to OpenClaw that runs on 10MB RAM versus OpenClaw's $599 Mac Mini requirement. The Go-based binary offers the same AI agent capabilities at 1/60th the hardware cost.
Apple M5 Max NPU Benchmarks 2x Faster Than Intel Panther Lake NPU in Parakeet v3 AI Inference Test
A leaked benchmark using the Parakeet v3 AI speech recognition model shows Apple's next-generation M5 Max Neural Processing Unit (NPU) delivering double the inference speed of Intel's competing Panther Lake NPU. This real-world test provides early performance data in the intensifying on-device AI hardware race.
Nvidia Claims MLPerf Inference v6.0 Records with 288-GPU Blackwell Ultra Systems, Highlights 2.7x Software Gains
MLCommons released MLPerf Inference v6.0 results, introducing multimodal and video model tests. Nvidia set records using 288-GPU Blackwell Ultra systems and achieved a 2.7x performance jump on DeepSeek-R1 via software optimizations alone.
TSMC 2nm Capacity Constraints Create Opening for Samsung in AI Chip Foundry Race
TSMC has reportedly hit a 'hard capacity wall' at its 2nm node, creating a strategic opportunity for Samsung Foundry to capture AI accelerator business from major clients like Nvidia and OpenAI. This bottleneck could reshape the competitive landscape for advanced semiconductor manufacturing.
Oracle Cuts 20% of Workforce to Fund AI Infrastructure Push, Shifting from Labor to Compute
Oracle is laying off 20% of its workforce to redirect capital toward massive AI infrastructure investments. The move signals a strategic pivot from traditional workforce costs to data center and compute spending.
Nvidia DLSS 4.5 Launches with Enhanced AI Frame Generation and Ray Reconstruction
Nvidia has released DLSS 4.5, a major update to its AI-powered upscaling technology featuring new frame generation modes and improved ray reconstruction. The update is available now for GeForce RTX 40 and 50 Series GPUs.
Ollama Now Supports Apple MLX Backend for Local LLM Inference on macOS
Ollama, the popular framework for running large language models locally, has added support for Apple's MLX framework as a backend. This enables more efficient execution of models like Llama 3.2 and Mistral on Apple Silicon Macs.