slm
12 articles about slm in AI news
Vibe Training: SLM Replaces LLM-as-a-Judge, 8x Faster, 50% Fewer Errors
Plurai introduces 'vibe training,' using adversarial agent swarms to distill a small language model (SLM) for evaluating and guarding production AI agents. The SLM outperforms standard LLM-as-a-judge setups with ~8x faster inference and ~50% fewer evaluation errors.
AI Writes New Virus DNA: Stanford and Arc Institute's DNA Language Model
A tweet reports that researchers fed a language model a DNA sequence and asked it to generate a new virus, which it did. This highlights both the power and risk of generative AI in synthetic biology.
Microsoft's Fairwater AI Data Center Launches Early, Boosts Azure Capacity
Microsoft has launched its Fairwater AI data center ahead of schedule. The facility adds significant high-performance computing capacity to Azure's AI infrastructure, crucial for training and running large models.
Claude-Obsidian Open-Source Plugin Aims to Automate Knowledge Management
A developer announced Claude-Obsidian, an open-source plugin that uses AI to autonomously file, cross-reference, and research within Obsidian, citing it as a reason to delete Notion AI.
DharmaOCR: New Small Language Models Set State-of-the-Art for Structured
A new arXiv preprint presents DharmaOCR, a pair of small language models (7B & 3B params) fine-tuned for structured OCR. They introduce a new benchmark and use Direct Preference Optimization to drastically reduce 'text degeneration'—a key cause of performance failures—while outputting structured JSON. The models claim superior accuracy and lower cost than proprietary APIs.
Pioneer Agent: A Closed-Loop System for Automating Small Language Model
Researchers present Pioneer Agent, a system that automates the adaptation of small language models to specific tasks. It handles data curation, failure diagnosis, and iterative training, showing significant performance gains in benchmarks and production-style deployments. This addresses a major engineering bottleneck for deploying efficient, specialized AI.
Meta's New Training Recipe: Small Models Should Learn from a Single Expert
Meta AI researchers propose a novel training recipe for small language models: instead of learning from many large 'expert' models simultaneously, they should be trained sequentially on one expert at a time. This method, detailed in a new paper, reportedly improves final model performance and training efficiency.
Ethan Mollick: Gemma 4 Impressive On-Device, But Agentic Workflows Doubted
Wharton professor Ethan Mollick finds Google's Gemma 4 powerful for on-device use but is skeptical about its ability to execute true agentic workflows, citing limitations in judgment and self-correction.
OpenClaw AI Agent Used for Stroller Repair, Sparking Debate on AI's Role in Human Connection
A viral tweet by George Pu highlights users employing AI agents like OpenClaw for mundane tasks like booking repairs and ranking friends, framing it as 'loneliness with a tech stack' rather than productivity.
MiRA Framework Boosts Gemma3-12B to 43% Success Rate on WebArena-Lite, Surpassing GPT-4 and WebRL
Researchers propose MiRA, a milestone-based RL framework that improves long-horizon planning in LLM agents. It boosts Gemma3-12B's web navigation success from 6.4% to 43%, outperforming GPT-4-Turbo (17.6%) and the previous SOTA WebRL (38.4%).
Fine-Tune Phi-3 Mini with Unsloth: A Practical Guide for Product Information Extraction
A technical tutorial demonstrates how to fine-tune Microsoft's compact Phi-3 Mini model using the Unsloth library for structured information extraction from product descriptions, all within a free Google Colab notebook.
Fine-Tuning Gemma 3 1B-IT for Financial Reasoning with QLoRA
A technical guide details using QLoRA and reasoning-augmented data to fine-tune Google's Gemma 3 1B-IT model for financial analysis. This demonstrates a method to specialize small language models for complex, domain-specific tasks.