apple
30 articles about apple in AI news
Developer Ranks NPU Model Compilation Ease: Apple 1st, AMD Last
Developer @mweinbach ranked the ease of using AI coding agents to compile ML models for NPUs. Apple's ecosystem was rated easiest, while AMD's tooling was ranked most difficult.
Gemma 4 Ported to MLX-Swift, Runs Locally on Apple Silicon
Google's Gemma 4 language model has been ported to the MLX-Swift framework by a community developer, making it available for local inference on Apple Silicon Macs and iOS devices through the LocallyAI app.
Apple M5 Max NPU Benchmarks 2x Faster Than Intel Panther Lake NPU in Parakeet v3 AI Inference Test
A leaked benchmark using the Parakeet v3 AI speech recognition model shows Apple's next-generation M5 Max Neural Processing Unit (NPU) delivering double the inference speed of Intel's competing Panther Lake NPU. This real-world test provides early performance data in the intensifying on-device AI hardware race.
Apple's Eddy Cue to Appear on TBPN Podcast for Company's 50th Anniversary
Apple's senior vice president of services, Eddy Cue, will appear live on the TBPN podcast today at 12:10 PM PT. The interview is part of Apple's 50th-anniversary commemorations.
Roboflow's RF-DETR Model Ported to Apple MLX, Enabling Real-Time On-Device Instance Segmentation
Roboflow's RF-DETR object detection model is now available on Apple's MLX framework, enabling real-time instance segmentation on Apple Silicon devices. This port unlocks new on-device visual analysis applications for robotics and augmented vision-language models.
Apple Removes AI Coding Apps Replit & Vibecode from App Store, Coinciding with Xcode AI Integration
Apple has removed AI-powered coding apps Replit and Vibecode from the App Store, reportedly for enabling app creation outside Apple's approval system. This coincides with Apple's recent integration of its own AI coding assistant into Xcode.
Ollama Now Supports Apple MLX Backend for Local LLM Inference on macOS
Ollama, the popular framework for running large language models locally, has added support for Apple's MLX framework as a backend. This enables more efficient execution of models like Llama 3.2 and Mistral on Apple Silicon Macs.
Apple Silicon Achieves Near-Lossless LLM Compression at 3.5 Bits-Per-Weight, Claims Independent Tester
Independent AI researcher Matthew Weinbach reports achieving near-lossless compression of large language models on Apple Silicon, storing models at 3.5 bits-per-weight while maintaining within 1-2% quality of bf16 precision.
Facebook's SAM 3 Vision Model Ported to Apple's MLX Framework, Enabling Real-Time Tracking on M3 Max
Facebook's Segment Anything Model 3 (SAM 3) has been ported to Apple's MLX framework, enabling real-time object tracking on an M3 Max MacBook Pro. This demonstrates efficient on-device execution of a foundational vision model without cloud dependency.
mlx-vlm v0.4.2 Adds SAM3, DOTS-MOCR Models and Critical Fixes for Vision-Language Inference on Apple Silicon
mlx-vlm v0.4.2 released with support for Meta's SAM3 segmentation model and DOTS-MOCR document OCR, plus fixes for Qwen3.5, LFM2-VL, and Magistral models. Enables efficient vision-language inference on Apple Silicon via MLX framework.
Apple's On-Device Reranking Model for Private Visual Search: A Technical Breakdown
Analysis of Apple's Enhanced Visual Search system that uses multimodal features, geo-signals, and index debiasing to identify landmarks entirely on-device. This represents a significant advancement in privacy-preserving AI for visual recognition.
Apple's Private Cloud Compute: Leak Suggests 4x M2 Ultra Cluster for On-Device AI Offload
A leak suggests Apple's Private Cloud Compute for AI may be built on clusters of four M2 Ultra chips, potentially offering high-performance, private server-side processing for iPhone AI tasks. This would mark Apple's strategic move into dedicated, privacy-focused AI infrastructure.
Apple Hires Former Google Exec Lilian Rincon as VP of AI Product Marketing
Apple has appointed Lilian Rincon, a former Google executive, as its Vice President of Product Marketing for Artificial Intelligence. This is a key strategic hire as Apple intensifies its push into consumer-facing AI products.
Qwen3-TTS Added to mlx-tune, Enabling Full Qwen Model Fine-Tuning on Apple Silicon Macs
The mlx-tune library now supports Qwen3-TTS, making the entire Qwen model stack—including the new text-to-speech model—fine-tunable on Apple Silicon Macs. This expands local AI development options for researchers and developers.
Apple iOS 27 to Introduce 'Extensions' for Siri, Allowing Users to Link to ChatGPT, Gemini, or Claude
Apple's iOS 27 will reportedly let users choose third-party AI chatbots like Google Gemini or Anthropic Claude to power Siri responses via a new 'Extensions' feature. This follows Apple's confirmed deal with Google to power its overhauled Siri, signaling a major shift from a closed to an open AI assistant ecosystem.
TurboQuant Ported to Apple MLX, Claims 75% Memory Reduction with Minimal Performance Loss
Developer Prince Canuma has successfully ported the TurboQuant quantization method to Apple's MLX framework, reporting a 75% reduction in memory usage with nearly no performance degradation for on-device AI models.
Apple Announces Plans to Increase US iPhone Parts Manufacturing, Continuing Supply Chain Diversification
Apple has announced plans to manufacture more iPhone components within the United States. This continues a multi-year strategy to diversify its supply chain away from concentrated geographic regions.
Apple Reportedly Gains Full Internal Access to Google's Gemini for On-Device Model Distillation
A report claims Apple's AI deal with Google includes full internal model access, enabling distillation of Gemini's reasoning into smaller, on-device models. This would allow Apple to build specialized, efficient AI without relying solely on cloud APIs.
Apple Siri Rebuilt as System-Wide AI Agent in iOS 27, Powered by Apple Foundation Models and Google Gemini
Apple is rebuilding Siri into a conversational system-wide AI agent with deep app integration and personal data access, launching in iOS 27. The overhaul includes a standalone app, web browsing, and writing tools, powered by Apple's models and a Google Gemini partnership.
Anthropic's Claude Reportedly Powers Apple's Internal Product Development Tools
Anthropic's AI models have reportedly become essential to Apple's internal operations, powering product development tools and contributing to the company's significant annual recurring revenue growth.
RunAnywhere's MetalRT Engine Delivers Breakthrough AI Performance on Apple Silicon
RunAnywhere has launched MetalRT, a proprietary GPU inference engine that dramatically accelerates on-device AI workloads on Apple Silicon. Their open-source RCLI tool demonstrates sub-200ms voice AI pipelines, outperforming existing solutions like llama.cpp and Apple's MLX.
AI Gold Rush Strains Apple Hardware: High-Memory Macs Sell Out as Local AI Agents Go Mainstream
A surge in demand for local AI development has created severe inventory shortages for high-memory Apple hardware. Mac Studio orders with 128GB or 512GB RAM face 6+ week delays as consumers buy up every available unit to run powerful AI agents like OpenClaw.
Apple's Neural Engine Jailbroken: Researchers Unlock Full Training Capabilities on M-Series Chips
Security researchers have reverse-engineered Apple's Neural Engine, bypassing private APIs to enable full neural network training directly on ANE hardware. This breakthrough unlocks 15.8 TFLOPS of compute previously restricted to inference-only operations across all M-series devices.
Open-Source Project Unlocks Apple's On-Device AI for Any Device on Your Network
Perspective Intelligence Web, an open-source project, enables any device with a browser to access Apple's powerful on-device AI models running locally on a Mac. This MIT-licensed solution addresses privacy concerns by keeping all processing on your private network while extending Apple Intelligence capabilities to Windows, Linux, Android, and Chromebook devices.
Apple's M5 Pro and Max: Fusion Architecture Redefines AI Computing on Silicon
Apple unveils M5 Pro and M5 Max chips with groundbreaking Fusion Architecture, merging two 3nm dies into a single SoC. The chips deliver up to 30% faster CPU performance and over 4x peak GPU compute for AI workloads compared to previous generations.
Apple's Neural Engine Jailbroken: Researchers Unlock On-Device AI Training Capabilities
A researcher has reverse-engineered Apple's private Neural Engine APIs to enable direct transformer training on M-series chips, bypassing CoreML restrictions. This breakthrough could enable battery-efficient local model training and fine-tuning without cloud dependency.
Apple Integrates AI Agents into Xcode 26.3, Transforming iOS Development
Apple has officially released Xcode 26.3 with native support for AI agents from Anthropic and OpenAI. This update enables developers to use Claude Agent and Codex directly within the IDE for autonomous app development tasks, marking a significant shift in Apple's developer ecosystem strategy.
Apple's 'Visual Intelligence' Vision: How AI-Powered Cameras Will Redefine Wearables
Apple is developing 'Visual Intelligence'—AI that interprets the physical world through cameras—as the foundation for its next generation of wearables, including smart glasses, advanced AirPods, and a camera-equipped pendant.
The AI Music Revolution: How Google and Apple Are Democratizing Music Creation
Google and Apple are integrating generative AI music features into their core platforms, allowing users to create custom 30-second tracks from text, photos, or video prompts. This move signals AI's transition from experimental tools to mainstream consumer applications.
mlx-vlm v0.4.4 Launches with Falcon-Perception 300M, TurboQuant Metal Kernels & 1.9x Decode Speedup
The mlx-vlm library v0.4.4 adds support for TII's Falcon-Perception 300M vision model and introduces TurboQuant Metal kernels, achieving up to 1.9x faster decoding with 89% KV cache savings on Apple Silicon.