Skip to content
gentic.news — AI News Intelligence Platform

slm

12 articles about slm in AI news

Vibe Training: SLM Replaces LLM-as-a-Judge, 8x Faster, 50% Fewer Errors

Plurai introduces 'vibe training,' using adversarial agent swarms to distill a small language model (SLM) for evaluating and guarding production AI agents. The SLM outperforms standard LLM-as-a-judge setups with ~8x faster inference and ~50% fewer evaluation errors.

82% relevant

AI Writes New Virus DNA: Stanford and Arc Institute's DNA Language Model

A tweet reports that researchers fed a language model a DNA sequence and asked it to generate a new virus, which it did. This highlights both the power and risk of generative AI in synthetic biology.

85% relevant

Microsoft's Fairwater AI Data Center Launches Early, Boosts Azure Capacity

Microsoft has launched its Fairwater AI data center ahead of schedule. The facility adds significant high-performance computing capacity to Azure's AI infrastructure, crucial for training and running large models.

92% relevant

Claude-Obsidian Open-Source Plugin Aims to Automate Knowledge Management

A developer announced Claude-Obsidian, an open-source plugin that uses AI to autonomously file, cross-reference, and research within Obsidian, citing it as a reason to delete Notion AI.

85% relevant

DharmaOCR: New Small Language Models Set State-of-the-Art for Structured

A new arXiv preprint presents DharmaOCR, a pair of small language models (7B & 3B params) fine-tuned for structured OCR. They introduce a new benchmark and use Direct Preference Optimization to drastically reduce 'text degeneration'—a key cause of performance failures—while outputting structured JSON. The models claim superior accuracy and lower cost than proprietary APIs.

72% relevant

Pioneer Agent: A Closed-Loop System for Automating Small Language Model

Researchers present Pioneer Agent, a system that automates the adaptation of small language models to specific tasks. It handles data curation, failure diagnosis, and iterative training, showing significant performance gains in benchmarks and production-style deployments. This addresses a major engineering bottleneck for deploying efficient, specialized AI.

74% relevant

Meta's New Training Recipe: Small Models Should Learn from a Single Expert

Meta AI researchers propose a novel training recipe for small language models: instead of learning from many large 'expert' models simultaneously, they should be trained sequentially on one expert at a time. This method, detailed in a new paper, reportedly improves final model performance and training efficiency.

85% relevant

Ethan Mollick: Gemma 4 Impressive On-Device, But Agentic Workflows Doubted

Wharton professor Ethan Mollick finds Google's Gemma 4 powerful for on-device use but is skeptical about its ability to execute true agentic workflows, citing limitations in judgment and self-correction.

75% relevant

OpenClaw AI Agent Used for Stroller Repair, Sparking Debate on AI's Role in Human Connection

A viral tweet by George Pu highlights users employing AI agents like OpenClaw for mundane tasks like booking repairs and ranking friends, framing it as 'loneliness with a tech stack' rather than productivity.

85% relevant

MiRA Framework Boosts Gemma3-12B to 43% Success Rate on WebArena-Lite, Surpassing GPT-4 and WebRL

Researchers propose MiRA, a milestone-based RL framework that improves long-horizon planning in LLM agents. It boosts Gemma3-12B's web navigation success from 6.4% to 43%, outperforming GPT-4-Turbo (17.6%) and the previous SOTA WebRL (38.4%).

77% relevant

Fine-Tune Phi-3 Mini with Unsloth: A Practical Guide for Product Information Extraction

A technical tutorial demonstrates how to fine-tune Microsoft's compact Phi-3 Mini model using the Unsloth library for structured information extraction from product descriptions, all within a free Google Colab notebook.

72% relevant

Fine-Tuning Gemma 3 1B-IT for Financial Reasoning with QLoRA

A technical guide details using QLoRA and reasoning-augmented data to fine-tune Google's Gemma 3 1B-IT model for financial analysis. This demonstrates a method to specialize small language models for complex, domain-specific tasks.

89% relevant