amd

30 articles about amd in AI news

AMD AI Director Reports Claude Code Quality Decline, Cites 234k Tool Calls

An AMD AI executive presented data from over 6,800 sessions showing Claude Code's performance has declined since early March, with rising instances of shallow reasoning and incomplete tasks. This raises significant trust issues for engineers using the model in complex development workflows.

85% relevant

MLPerf 6.0: NVIDIA Sweeps New Benchmarks, AMD MI355X Within 30% on Select Tests

MLPerf 6.0 results show NVIDIA winning every new benchmark, with its GB300 NVL72 system achieving nearly 3x more throughput than six months ago. AMD's MI355X showed progress, coming within 10-30% on select single-node tests but skipping most new benchmarks.

85% relevant

Developer Ranks NPU Model Compilation Ease: Apple 1st, AMD Last

Developer @mweinbach ranked the ease of using AI coding agents to compile ML models for NPUs. Apple's ecosystem was rated easiest, while AMD's tooling was ranked most difficult.

75% relevant

Meta's $100B AMD Gamble: The AI Chip War Enters Its Most Strategic Phase

Meta has secured a landmark deal to purchase up to $100 billion worth of AMD AI chips, receiving a massive stock warrant in return. This unprecedented agreement signals Meta's aggressive push to diversify its AI infrastructure beyond Nvidia while pursuing ambitious 'personal superintelligence' goals.

90% relevant

Meta's $100 Billion AMD Bet: The AI Infrastructure Arms Race Reaches New Heights

Meta has reportedly signed a staggering $100 billion agreement with AMD to secure 6GW of data center capacity, signaling an unprecedented commitment to AI infrastructure. The timing—just before NVIDIA's quarterly results—highlights intensifying competition for computing resources essential for next-generation AI models.

95% relevant

NVIDIA's DreamDojo: Teaching Robots to 'Dream' in Pixels with 44,000 Hours of Human Experience

NVIDIA has open-sourced DreamDojo, a revolutionary robot world model trained on 44,711 hours of real-world human video. Instead of relying on physics engines, it predicts action outcomes directly in pixel space, potentially accelerating robotics development by orders of magnitude.

85% relevant

DeepSeek V4 Launch Signals China's Strategic Shift in AI Chip Independence

DeepSeek's upcoming V4 multimodal model prioritizes domestic chip partners Huawei and Cambricon over NVIDIA and AMD, marking a significant move toward Chinese AI self-sufficiency amid ongoing U.S. export restrictions.

78% relevant

Anthropic Considers Custom AI Chips, Following Google & OpenAI

Anthropic is reportedly considering developing custom AI chips, a strategic move to gain control over its compute infrastructure and reduce costs. This follows similar initiatives by Google, Amazon, and OpenAI.

85% relevant

Intel & Google Announce Multiyear AI & Cloud Infrastructure Partnership

Intel and Google have announced a multiyear strategic collaboration to advance AI and cloud infrastructure, focusing on optimizing Google Cloud for Intel's Xeon processors, Gaudi AI accelerators, and future chips.

85% relevant

Intel, SambaNova Blueprint Pairs GPUs for AI Prefill, RDUs for Decoding

Intel and SambaNova Systems have outlined a new inference architecture for agentic AI workloads. It splits tasks between GPUs for 'prefill' and SambaNova's Reconfigurable Dataflow Units (RDUs) for high-throughput token generation.

85% relevant

Broadcom to Manufacture Google TPU Chips in Foundry Partnership

Google has licensed its Tensor Processing Unit (TPU) intellectual property to Broadcom for chip fabrication. This allows Google to earn from its IP while Broadcom manages the complex hardware build and networking integration.

85% relevant

Dell XPS 14 with Core Ultra X7 Outlasts Snapdragon X Elite in Battery Test

A new battery test shows the Dell XPS 14 with Intel Core Ultra X7 lasts 11.7 hours, beating the 11.3 hours of Microsoft's ARM-based Surface Laptop 15 with Snapdragon X Elite. This is a significant win for Intel's latest chip in the critical mobile performance metric.

85% relevant

Qualcomm X2 Elite Matches Apple M5 in Efficiency Test

In a mixed-use laptop test simulating office work, Qualcomm's Snapdragon X2 Elite system-on-chip matched the power efficiency of Apple's latest M5 chip. This marks a significant milestone for Windows on Arm in its competition with Apple Silicon.

75% relevant

Terafab's 1GW AI Compute Goal Requires Massive Fab Capacity

Analysis of Terafab's stated goals shows that achieving 1GW of AI compute would require approximately 190,000 wafer starts per month across logic and memory. This underscores the unprecedented scale of semiconductor manufacturing needed for future AI infrastructure.

85% relevant

Microsoft's BitNet Enables 100B-Parameter LLMs on CPU, Cuts Energy 82%

Microsoft Research's BitNet project demonstrates 1-bit LLMs with 100B parameters that run efficiently on CPUs, using 82% less energy while maintaining performance, challenging the need for GPUs in local deployment.

95% relevant

Intel Joins SpaceX, xAI, Tesla in 'Terafab' Chip Project

Intel announced it is joining the 'Terafab' project alongside SpaceX, xAI, and Tesla. The collaboration aims to refactor silicon fab technology, likely to support the massive compute demands of AI and aerospace.

85% relevant

FDA-Designated AI 'Vox' Detects Heart Failure from 5-Second Voice Clip

An AI tool named Vox can detect signs of worsening heart failure from a 5-second patient voice clip. It's trained on >3M voice samples and backed by five clinical trials, targeting a condition affecting 64M people globally.

95% relevant

McKinsey: AI Infrastructure Value Creation Outpaces Business Capture

McKinsey's latest analysis indicates the pace of value creation from AI infrastructure is exceeding the rate at which most businesses are capturing it, highlighting a growing implementation deficit.

75% relevant

US Data Center Power Demand Hits 15 GW, Grid Constraints Emerge

US data center power demand reached 15 gigawatts in 2023, up from 11 GW in 2022. This rapid growth highlights a widening bottleneck: compute infrastructure is scaling faster than power delivery systems can support.

75% relevant

OpenAI, Anthropic Forecast $121B Compute Burn, Revealing AI's True Cost

Internal forecasts from OpenAI and Anthropic reveal the core challenge of modern AI has shifted from selling the technology to financing the immense compute required for training and inference, with OpenAI projecting $121B in compute spending for 2028.

99% relevant

Neuromorphic Computing Patents Surge 401% in 2025, Hits 596 by 2026

Patent filings for neuromorphic computing—hardware that mimics the brain's architecture—surged 401% in 2025, reaching 596 by early 2026. This indicates the technology is transitioning from lab prototypes to commercial products.

87% relevant

Chamath Palihapitiya: AI's Biggest Profits Won't Go to Model Makers

VC Chamath Palihapitiya posits that the greatest financial winners in AI will be application builders with unique distribution, not the foundational model creators, drawing a parallel to refrigeration and Coca-Cola.

75% relevant

X Post Reveals Audible Quality Differences in GPU vs. NPU AI Inference

A developer demonstrated audible quality differences in AI text-to-speech output when run on GPU, CPU, and NPU hardware, highlighting a key efficiency vs. fidelity trade-off for on-device AI.

75% relevant

PicoClaw: $10 RISC-V AI Agent Challenges OpenClaw's $599 Mac Mini Requirement

Developers have launched PicoClaw, a $10 RISC-V alternative to OpenClaw that runs on 10MB RAM versus OpenClaw's $599 Mac Mini requirement. The Go-based binary offers the same AI agent capabilities at 1/60th the hardware cost.

87% relevant

Apple M5 Max NPU Benchmarks 2x Faster Than Intel Panther Lake NPU in Parakeet v3 AI Inference Test

A leaked benchmark using the Parakeet v3 AI speech recognition model shows Apple's next-generation M5 Max Neural Processing Unit (NPU) delivering double the inference speed of Intel's competing Panther Lake NPU. This real-world test provides early performance data in the intensifying on-device AI hardware race.

85% relevant

Nvidia Claims MLPerf Inference v6.0 Records with 288-GPU Blackwell Ultra Systems, Highlights 2.7x Software Gains

MLCommons released MLPerf Inference v6.0 results, introducing multimodal and video model tests. Nvidia set records using 288-GPU Blackwell Ultra systems and achieved a 2.7x performance jump on DeepSeek-R1 via software optimizations alone.

95% relevant

TSMC 2nm Capacity Constraints Create Opening for Samsung in AI Chip Foundry Race

TSMC has reportedly hit a 'hard capacity wall' at its 2nm node, creating a strategic opportunity for Samsung Foundry to capture AI accelerator business from major clients like Nvidia and OpenAI. This bottleneck could reshape the competitive landscape for advanced semiconductor manufacturing.

85% relevant

Oracle Cuts 20% of Workforce to Fund AI Infrastructure Push, Shifting from Labor to Compute

Oracle is laying off 20% of its workforce to redirect capital toward massive AI infrastructure investments. The move signals a strategic pivot from traditional workforce costs to data center and compute spending.

97% relevant

Nvidia DLSS 4.5 Launches with Enhanced AI Frame Generation and Ray Reconstruction

Nvidia has released DLSS 4.5, a major update to its AI-powered upscaling technology featuring new frame generation modes and improved ray reconstruction. The update is available now for GeForce RTX 40 and 50 Series GPUs.

85% relevant

Ollama Now Supports Apple MLX Backend for Local LLM Inference on macOS

Ollama, the popular framework for running large language models locally, has added support for Apple's MLX framework as a backend. This enables more efficient execution of models like Llama 3.2 and Mistral on Apple Silicon Macs.

85% relevant