representation learning
30 articles about representation learning in AI news
LLM Schema-Adaptive Method Enables Zero-Shot EHR Transfer
Researchers propose Schema-Adaptive Tabular Representation Learning, an LLM-driven method that transforms structured variables into semantic statements. It enables zero-shot alignment across unseen EHR schemas and outperforms clinical baselines, including neurologists, on dementia diagnosis tasks.
Uber Eats Details Production System for Multilingual Semantic Search Across Stores, Dishes, and Items
Uber Eats engineers published a paper detailing their production semantic retrieval system that unifies search across stores, dishes, and grocery items using a fine-tuned Qwen2 model. The system leverages Matryoshka Representation Learning to serve multiple embedding sizes and shows substantial recall gains across six markets.
CoRe Framework Integrates Equivariant Contrastive Learning for Medical Image Registration, Surpassing Baseline Methods
Researchers propose CoRe, a medical image registration framework that jointly optimizes an equivariant contrastive learning objective with the registration task. The method learns deformation-invariant feature representations, improving performance on abdominal and thoracic registration tasks.
SPREAD Framework Solves AI's 'Catastrophic Forgetting' Problem in Lifelong Learning
Researchers have developed SPREAD, a new AI framework that preserves learned skills across sequential tasks by aligning policy representations in low-rank subspaces. This breakthrough addresses catastrophic forgetting in lifelong imitation learning, enabling more stable and robust AI agents.
MVCrec: A New Multi-View Contrastive Learning Framework for Sequential
Researchers propose MVCrec, a framework that applies multi-view contrastive learning between sequential (ID-based) and graph-based views of user interaction data to improve recommendation accuracy. It outperforms 11 leading models, showing significant gains in key metrics.
Google's RT-X Project Establishes New Robot Learning Standard
Google's RT-X project has established a new standard for robot learning by creating a unified dataset of detailed human demonstrations across 22 institutions and 30+ robot types. This enables large-scale cross-robot training previously impossible with fragmented data.
New Relative Contrastive Learning Framework Boosts Sequential Recommendation Accuracy by 4.88%
A new arXiv paper introduces Relative Contrastive Learning (RCL) for sequential recommendation. It solves a data scarcity problem in prior methods by using similar user interaction sequences as additional training signals, leading to significant accuracy improvements.
The Future of Production ML Is an 'Ugly Hybrid' of Deep Learning, Classic ML, and Rules
A technical article argues that the most effective production machine learning systems are not pure deep learning or classic ML, but pragmatic hybrids combining embeddings, boosted trees, rules, and human review. This reflects a maturing, engineering-first approach to deploying AI.
Meta's V-JEPA 2.1 Achieves +20% Robotic Grasp Success with Dense Feature Learning from 1M+ Hours of Video
Meta researchers released V-JEPA 2.1, a video self-supervised learning model that learns dense spatial-temporal features from over 1 million hours of video. The approach improves robotic grasp success by ~20% over previous methods by forcing the model to understand precise object positions and movements.
Building a Smart Learning Path Recommendation System Using Graph Neural Networks
A technical article outlines how to build a learning path recommendation system using Graph Neural Networks (GNNs). It details constructing a knowledge graph and applying GNNs for personalized course sequencing, a method with clear parallels to retail product discovery.
FedShare: A New Framework for Federated Recommendation with Personalized Data Sharing and Unlearning
Researchers propose FedShare, a federated learning framework for recommender systems that allows users to dynamically share data for better performance and request its removal via efficient 'unlearning', addressing a key privacy-performance trade-off.
Hierarchical AI Breakthrough: Meta-Reinforcement Learning Unlocks Complex Task Mastery Through Skill-Based Curriculum
Researchers have developed a novel multi-level meta-reinforcement learning framework that compresses complex decision-making problems into hierarchical structures, enabling AI to master intricate tasks through skill-based curriculum learning. This approach reduces computational complexity while improving transfer learning across different problems.
ATLAS: Pioneering Lifelong Learning for AI That Sees and Hears
Researchers introduce the first continual learning benchmark for audio-visual segmentation, addressing how AI systems can adapt to evolving real-world environments without forgetting previous knowledge. The ATLAS framework uses audio-guided conditioning and low-rank anchoring to maintain performance across dynamic scenarios.
Reinforcement Learning Ushers in New Era of Autonomous Knowledge Agents
Researchers are developing knowledge agents powered by reinforcement learning that can autonomously gather, process, and apply information. These systems represent a significant evolution beyond traditional language models toward more independent problem-solving capabilities.
AI Researchers Crack the Delay Problem: New Algorithm Achieves Optimal Performance in Real-World Reinforcement Learning
Researchers have developed a minimax optimal algorithm for reinforcement learning with delayed state observations, achieving provably optimal regret bounds. This breakthrough addresses a fundamental challenge in real-world AI systems where sensors and processing create unavoidable latency.
Beyond Homogenization: How Expert Divergence Learning Unlocks MoE's True Potential
Researchers have developed Expert Divergence Learning, a novel pre-training strategy that combats expert homogenization in Mixture-of-Experts language models. By encouraging functional specialization through domain-aware routing, the method improves performance across benchmarks with minimal computational overhead.
Beyond Flat Space: How Hyperbolic Geometry Solves AI's Few-Shot Learning Bottleneck
Researchers propose Hyperbolic Flow Matching (HFM), a novel approach using hyperbolic geometry to dramatically improve few-shot learning. By leveraging the exponential expansion of Lorentz manifolds, HFM prevents feature entanglement that plagues traditional Euclidean methods, achieving state-of-the-art results across 11 benchmarks.
MemRerank: A Reinforcement Learning Framework for Distilling Purchase History into Personalized Product Reranking
Researchers propose MemRerank, a framework that uses RL to distill noisy user purchase histories into concise 'preference memory' for LLM-based shopping agents. It improves personalized product reranking accuracy by up to +10.61 points versus raw-history baselines.
Google DeepMind's 'Learning Through Conversation' Paper Shows LLMs Can Improve with Real-Time Feedback
Google DeepMind researchers have published a paper demonstrating that large language models can be trained to learn and improve their responses during a conversation by incorporating user feedback, moving beyond static pre-training.
Aligning Language Models from User Interactions: A Self-Distillation Method for Continuous Learning
Researchers propose a method to align LLMs using raw, multi-turn user conversations. By applying self-distillation on follow-up messages, models improve without explicit feedback, enabling personalization and continual adaptation from deployment data.
The Unlearning Illusion: New Research Exposes Critical Flaws in AI Memory Removal
Researchers reveal that current methods for making AI models 'forget' information are surprisingly fragile. A new dynamic testing framework shows that simple query modifications can recover supposedly erased knowledge, exposing significant safety and compliance risks.
Machine Learning Adventures: Teaching a Recommender System to Understand Outfits
A technical walkthrough of building an outfit-aware recommender system for a clothing marketplace. The article details the data pipeline, model architecture, and challenges of moving from single-item to outfit-level recommendations.
VGGT-Det: How AI Is Learning to See in 3D Without Camera Calibration
Researchers have developed VGGT-Det, a breakthrough framework for multi-view 3D object detection that works without calibrated camera poses. The system mines internal geometric priors through attention mechanisms, outperforming traditional methods in indoor environments.
Cross-View AI System Masters Object Matching Without Supervision
A novel CVPR 2026 framework achieves robust object correspondence between first-person and third-person views using cycle-consistent mask prediction, eliminating the need for costly manual annotations while learning view-invariant representations.
AFMRL: Using MLLMs to Generate Attributes for Better Product Retrieval in
AFMRL uses MLLMs to generate product attributes, then uses those attributes to train better multimodal representations for e-commerce retrieval. Achieves SOTA on large-scale datasets.
VoteGCL: A Novel LLM-Augmented Framework to Combat Data Sparsity in
A new paper introduces VoteGCL, a framework that uses few-shot LLM prompting and majority voting to create high-confidence synthetic data for graph-based recommendation systems. It integrates this data via graph contrastive learning to improve accuracy and mitigate bias, outperforming existing baselines.
LeWorldModel Solves JEPA Collapse with 15M Params, Trains on Single GPU
Researchers published LeWorldModel, solving the representation collapse problem in Yann LeCun's JEPA architecture. The 15M-parameter model trains on a single GPU and demonstrates intrinsic physics understanding.
IPCCF: A New Graph-Based Approach to Disentangle User Intent for Better
A new research paper introduces Intent Propagation Contrastive Collaborative Filtering (IPCCF), a method designed to improve recommendation systems by more accurately disentangling the underlying intents behind user-item interactions. It addresses limitations in existing methods by incorporating broader graph structure and using contrastive learning for direct supervision, showing superior performance in experiments.
DUET: A New LLM-Based Recommender That Generates Paired User-Item Profiles
A new research paper introduces DUET, an interaction-aware profile generator for recommendation systems. Instead of using dense vectors or independent text descriptions, it jointly creates semantically consistent user and item profiles conditioned on their interaction history, optimizing them with reinforcement learning for better performance.
Dual-Enhancement Product Bundling
Researchers propose a dual-enhancement method for product bundling that integrates interactive graph learning with LLM-based semantic understanding. Their graph-to-text paradigm with Dynamic Concept Binding Mechanism addresses cold-start problems and graph comprehension limitations, showing significant performance gains on benchmarks.