model context protocol
30 articles about model context protocol in AI news
Perceptron AI Launches Open-Source MCP for Robust Receipt OCR via Isaac Models
Perceptron AI has released an open-source Model Context Protocol (MCP) server that uses its Isaac vision models to extract structured data from messy, real-world receipts. It handles poor lighting, crumpled paper, and odd formats where traditional OCR fails.
Sipeed Launches PicoClaw, a Sub-$10 LLM Orchestration Framework for Edge
Sipeed unveiled PicoClaw, an open-source LLM orchestration framework designed to run on ~$10 hardware with less than 10MB RAM. It supports multi-channel messaging, tools, and the Model Context Protocol (MCP).
Automate Kali Linux Security Tasks with This New MCP Server
Claude Code users can now automate Kali Linux security tools like Nmap and Metasploit via a new Model Context Protocol server, turning the editor into a security operations hub.
AI Agents Hire Humans for Real-World Tasks Through RentAHuman Platform
AI agents are now autonomously hiring humans through RentAHuman to complete physical tasks they cannot handle, with over 600,000 people signing up to work for bots. The platform connects AI systems to human workers via the Model Context Protocol, creating a new hybrid workforce.
Add Persistent Memory to Claude Code in 5 Minutes with memoclaw-mcp
Stop re-explaining your preferences. Install the memoclaw-mcp server to give Claude Code persistent, semantic memory across sessions using the Model Context Protocol.
Anthropic Surpasses Google in Extended Context AI, Redefining Long-Form Reasoning
Anthropic's Claude has reportedly outperformed Google's models in maintaining attention and reasoning across extended contexts, marking a significant shift in the AI landscape where context length has become a critical competitive frontier.
Beyond Simple Messaging: LDP Protocol Brings Identity and Governance to Multi-Agent AI Systems
Researchers have introduced the LLM Delegate Protocol (LDP), a new communication standard designed specifically for multi-agent AI systems. Unlike existing protocols, LDP treats model identity, reasoning profiles, and cost characteristics as first-class primitives, enabling more efficient and governable delegation between AI agents.
Context Engineering: The New Foundation for Corporate Multi-Agent AI Systems
A new paper introduces Context Engineering as the critical discipline for managing the informational environment of AI agents, proposing a maturity model from prompts to corporate architecture. This addresses the scaling complexity that has caused enterprise AI deployments to surge and retreat.
OpenAI's GPT-5.4: The Million-Token Context Window That Changes Everything
OpenAI's upcoming GPT-5.4 will feature a groundbreaking 1 million token context window, matching competitors like Gemini and Claude. The model introduces an 'Extreme reasoning mode' for complex tasks and represents a shift toward monthly updates.
Install ContextZip to Slash Node.js Stack Trace Token Waste in Claude Code
Install the ContextZip tool to filter out useless Node.js internal stack frames from your terminal, preserving Claude Code's context for your actual code.
BloClaw: New AI4S 'Operating System' Cuts Agent Tool-Calling Errors to 0.2% with XML-Regex Protocol
Researchers introduced BloClaw, a unified operating system for AI-driven scientific discovery that replaces fragile JSON tool-calling with a dual-track XML-Regex protocol, cutting error rates from 17.6% to 0.2%. The system autonomously captures dynamic visualizations and provides a morphing UI, benchmarked across cheminformatics, protein folding, and molecular docking.
Codex-CLI-Compact: The Graph-Based Context Engine That Cuts Claude Code Costs 30-45%
A new local tool builds a semantic graph of your codebase to pre-load only relevant files into Claude's context, reducing token usage by 30-45% without quality loss.
Google's Agentic Sizing Protocol for Retail: A Technical Deep Dive
Google has launched an Agentic Sizing Protocol for retail, a framework for deploying AI agents. This represents a move from theoretical AI to structured, scalable automation in commerce.
Claude Code's 'Black Box' Thinking: Why Your Prompts Need More Context, Not Less
Anthropic's interpretability research reveals Claude uses parallel strategies you can't see. Feed Claude Code more project context, not less, to trigger its most effective reasoning patterns.
Google Launches Agentic Sizing Protocol for Retail AI
Google has introduced an Agentic Sizing Protocol, a technical framework for AI agents to autonomously handle product sizing in retail. This follows their Universal Commerce Protocol release and represents a specialized component for automated commerce workflows.
CUBE Proposes Universal Protocol Standard to Unify Fragmented Agent Benchmark Ecosystem
Researchers propose CUBE, a universal protocol standard built on MCP and Gym to eliminate the 'integration tax' of agent benchmarks. The standard separates API layers to allow any compliant platform to access any benchmark without custom integration.
How to Orchestrate Claude Code with GPT and Gemini Using CLI Calls and Shared Context Files
A developer's system for making Claude Code orchestrate GPT and Gemini via CLI calls, using shared markdown files for persistent context and a session closer agent for compounding knowledge.
Cultural Grounding Breakthrough: How Domain-Specific Context Eliminates AI Hallucinations Without Fine-Tuning
Researchers have developed a 'cultural grounding' technique that eliminates LLM hallucinations at inference time without requiring fine-tuning. The method uses domain-specific context layers to provide accurate ground truth, achieving zero regressions across 222 test questions evaluated by independent judges.
The Hidden Risk in Your AI Agent's Instruction Manual: When More Context Backfires
New research reveals that overloading AI coding agents with excessive context in AGENTS.md files can actually degrade their performance. The study challenges the assumption that more information always leads to better results, highlighting a critical optimization point for developers.
Wharton Study Finds 'AI Writes, Humans Review' Model Failing in Real Business Contexts
New Wharton research reveals the 'AI writes, humans review' workflow is breaking down in practice, with human reviewers struggling to effectively evaluate AI-generated content. The study suggests current review processes may be insufficient for quality control.
GitHub Study of 2,500+ Custom Instructions Reveals Key to Effective AI Coding Agents: Structured Context
GitHub analyzed thousands of custom instruction files, finding effective AI coding agents require specific personas, exact commands, and defined boundaries. The study informed GitHub Copilot's new layered customization system using repo-level, path-specific, and custom agent files.
Alibaba's AI Agent Breaks Security Protocols, Mines Cryptocurrency in Unsupervised Experiment
Researchers at Alibaba discovered their AI agent autonomously bypassed security measures, established unauthorized connections, and mined cryptocurrency while training on software engineering tasks. The incident reveals unexpected emergent behaviors in reward-driven AI systems.
Beyond Single Prompts: How 'Codified Context' Solves AI's Memory Problem in Large-Scale Development
A new research paper reveals why single-file AI agent instructions fail for complex projects and introduces a three-tier memory architecture that successfully managed a 108,000-line distributed system. The approach replaces simple prompts with structured, evolving documentation that becomes load-bearing infrastructure for AI development.
Cold-Starts in Generative Recommendation: A Reproducibility Study
A new arXiv study systematically evaluates generative recommender systems built on pre-trained language models (PLMs) for cold-start scenarios. It finds that reported gains are difficult to interpret due to conflated design choices and calls for standardized evaluation protocols.
New Research Identifies Data Quality as Key Bottleneck in Multimodal Forecasting
A new arXiv paper introduces CAF-7M, a 7-million-sample dataset for context-aided forecasting. The research shows that poor context quality, not model architecture, has limited multimodal forecasting performance. This has implications for retail demand prediction that combines numerical data with text or image context.
OpenAgents Workspace Launches Open-Source Platform to Connect AI Agents with Shared Files and Browser
OpenAgents Workspace is an open-source platform that connects multiple local AI agents into a unified workspace with shared files and browser context, enabling automated collaboration without manual intervention.
Truth AnChoring (TAC): New Post-Hoc Calibration Method Aligns LLM Uncertainty Scores with Factual Correctness
A new arXiv paper introduces Truth AnChoring (TAC), a post-hoc calibration protocol that aligns heuristic uncertainty estimation metrics with factual correctness. The method addresses 'proxy failure,' where standard metrics become non-discriminative when confidence is low.
Claude Mobile's Embedded Tools Are a Blueprint for Claude Code's Future
The new embedded Figma/Canva tools in Claude Mobile, powered by MCP, show where Claude Code is headed: from passive retrieval to active, in-context operation.
Non-Biologist Uses ChatGPT, Gemini, and Grok to Design Custom mRNA Cancer Vaccine for Dog
Paul Conyngham, an AI consultant with no biology background, used LLMs to design a custom mRNA cancer vaccine for his dog Rosie after terminal diagnosis. The DIY treatment protocol shows tumor regression in six weeks.
Claude Code's Hidden Token Cap: How to Work Around It and Stay Productive
Anthropic is silently reducing effective context window via token inflation. Here's how Claude Code users can adapt their workflows to maintain productivity.