Skip to content
gentic.news — AI News Intelligence Platform

motion generation

30 articles about motion generation in AI news

LLaMo: The First Truly Unified Motion-Language AI Model That Understands and Generates Human Movement

Researchers have developed LLaMo, a groundbreaking AI model that unifies motion understanding and generation with language capabilities. Unlike previous approaches that suffered from catastrophic forgetting, LLaMo preserves linguistic knowledge while achieving real-time motion generation at over 30 FPS.

75% relevant

NVIDIA DLSS 5 Leak Suggests AI Frame Generation Without Motion Vectors

A leaked NVIDIA roadmap slide suggests DLSS 5 will use a new 'AI Frame Generation' technique that does not rely on traditional motion vectors, potentially simplifying game integration. The feature is slated for a 2026 release.

85% relevant

Kling AI 3.0 Arrives with Breakthrough Motion Control for Video Generation

Kling AI has launched version 3.0 featuring advanced motion control capabilities, representing a significant leap in AI-generated video technology. The update promises more precise manipulation of movement within AI-created videos.

85% relevant

NVIDIA Open-Sources Motion Diffusion Model for Humanoid Robots

NVIDIA open-sourced Kimono, a motion diffusion model for humanoid robots, trained on 700 hours of motion capture data. It generates 3D human and robot motions from text prompts, supports keyframe and end-effector control, and runs on Unitree G1.

85% relevant

India's Human Motion Farms Train Humanoid Robots with First-Person Hand Data

Labs in India are capturing detailed human motion data—focusing on grip, force, and error recovery—to train AI models for humanoid robots. This addresses the critical bottleneck of acquiring physical intelligence data for robotics.

89% relevant

Anthropic Paper: 'Emotion Concepts and their Function in LLMs' Published

Anthropic has released a new research paper titled 'Emotion Concepts and their Function in LLMs.' The work investigates the role and representation of emotional concepts within large language model architectures.

95% relevant

Bones Studio Demos Motion-Capture-to-Robot Pipeline for Home Tasks

Bones Studio released a demo showing its 'Captured → Labeled → Transferred' pipeline. It uses optical motion capture to record human tasks, then transfers the data for a humanoid robot to replicate the actions in simulation.

85% relevant

Zilan Lin on AI-Driven Motion Design and Redefining Luxury Visuals for the Gen Z Era

An interview with creative director Zilan Lin explores how AI-powered motion design tools are being used to create more dynamic, authentic, and culturally relevant visual content for luxury brands targeting Gen Z consumers.

72% relevant

E-STEER: New Framework Embeds Emotion in LLM Hidden States, Shows Non-Monotonic Impact on Reasoning and Safety

A new arXiv paper introduces E-STEER, an interpretable framework for embedding emotion as a controllable variable in LLM hidden states. Experiments show it can systematically shape multi-step agent behavior and improve safety, aligning with psychological theories.

75% relevant

Nvidia DLSS 4.5 Launches with Enhanced AI Frame Generation and Ray Reconstruction

Nvidia has released DLSS 4.5, a major update to its AI-powered upscaling technology featuring new frame generation modes and improved ray reconstruction. The update is available now for GeForce RTX 40 and 50 Series GPUs.

85% relevant

UiPath Launches AI Agents for Retail Pricing, Promotions, and Stock Management

UiPath has announced new AI agents designed to autonomously handle core retail operations: dynamic pricing, promotional planning, and inventory gap resolution. This represents a significant move by a major automation player into agentic AI for retail.

95% relevant

New Research Improves Text-to-3D Motion Retrieval with Interpretable Fine-Grained Alignment

Researchers propose a novel method for retrieving 3D human motion sequences from text descriptions using joint-angle motion images and token-patch interaction. It outperforms state-of-the-art methods on standard benchmarks while offering interpretable correspondences.

75% relevant

PAI Emerges as Potential Game-Changer in AI Video Generation Landscape

PAI has launched publicly, offering a new approach to AI video generation that prioritizes character consistency and narrative coherence. Early testing suggests it may address key limitations of current video AI systems.

85% relevant

AI Video Generation Reaches New Milestone: Kling AI 5.3 Launches with Enhanced Capabilities

The latest version of Kling AI, version 5.3, has officially launched, marking another advancement in AI-powered video generation technology. Early adopters are already sharing YouTube demonstrations showcasing improved capabilities.

85% relevant

Beyond Logic: How EMO-R3 Teaches AI to Reason About Human Emotions

Researchers have developed EMO-R3, a novel framework that enhances emotional reasoning in multimodal AI systems. Using reflective reinforcement learning, it enables AI to better understand and interpret human emotions in visual contexts, addressing a critical gap in current models.

80% relevant

R1's Real-Time World Model: The Paradigm Shift from Video Generation to World Generation

Rabbit's R1 introduces a real-time world model that continuously generates evolving environments rather than static video frames. This represents a fundamental shift from passive content creation to interactive world simulation, enabling seamless AI interactions without waiting or regeneration cycles.

85% relevant

The Dawn of Emotional AI Avatars: How Synthetic Humans Are Redefining Digital Interaction

New AI avatar technology creates emotionally responsive digital humans with realistic facial expressions, enabling natural conversations that could transform customer service, education, and social interaction.

85% relevant

From Static Suggestions to Dynamic Dialogue: The Next Generation of AI Recommendations for Luxury Retail

The AI recommendation market is projected to reach $34.4B by 2033, driven by advanced models like Google's Gemini that enable conversational, multi-modal personalization. For luxury brands, this means moving beyond basic 'customers also bought' to rich, contextual clienteling that understands taste, occasion, and brand heritage.

90% relevant

Nano Banana 2 Emerges: The Next Generation of AI-Powered Creative Tools

The AI creative community is abuzz with the apparent rollout of Nano Banana 2, a mysterious new tool that appears to build upon its predecessor's capabilities for generating and manipulating digital content through advanced machine learning models.

85% relevant

Lyria 3 Breaks Language Barriers: AI Music Generation Goes Truly Global

Google's Lyria 3 AI music model demonstrates unprecedented multilingual capabilities, generating authentic songs in languages beyond English. This breakthrough suggests AI music tools may soon serve global creative communities equally.

85% relevant

Dreamina Seedance 2.0 Early Access Review: AI Video Tool Adds Scene Direction Controls

An early tester reports that Dreamina Seedance 2.0 provides unprecedented control over AI-generated video, including camera motion, pacing, and visual consistency. The tool shifts from simple clip generation toward AI-native scene direction.

85% relevant

Kling AI Video Platform Goes Global: How 3.0 Release Redefines Accessible Cinematic AI

Kling AI has launched its 3.0 platform worldwide, offering 1080p cinematic video generation and advanced motion control. This marks a significant step toward professional-grade AI video tools becoming accessible to global creators.

85% relevant

The Cinematic AI Revolution: How Sora 2 Pro, Veo 3.1, and Kling 2.6 Are Democratizing Hollywood-Quality Video Production

OpenAI's Sora 2 Pro, Google's Veo 3.1, and Kling 2.6 represent a quantum leap in AI video generation, transforming text and images into cinematic-quality videos in minutes. These models offer Hollywood-level production values with smooth motion and clean lip sync, available through subscription models without per-video fees.

85% relevant

OpenAI Teases 'Not a Screenshot' AI Video Model

OpenAI posted a cryptic tweet stating 'This is not a screenshot' with a video link, strongly hinting at a new AI video generation model. This marks a direct move into a space currently led by rivals like Runway and Pika.

85% relevant

RAG vs Fine-Tuning vs Prompt Engineering

A technical blog clarifies that Retrieval-Augmented Generation (RAG), fine-tuning, and prompt engineering should be viewed as a layered stack, not mutually exclusive options. It provides a decision framework for when to use each technique based on specific needs like data freshness, task specificity, and cost.

90% relevant

LLM-HYPER: A Training-Free Framework for Cold-Start Ad CTR Prediction

A new arXiv paper introduces LLM-HYPER, a framework that treats large language models as hypernetworks to generate parameters for click-through rate estimators in a training-free manner. It uses multimodal ad content and few-shot prompting to infer feature weights, drastically reducing the cold-start period for new promotional ads and has been deployed on a major U.S. e-commerce platform.

96% relevant

ByteDance's OmniShow Unifies Text, Image, Audio, Pose for Video Gen

ByteDance introduced OmniShow, a unified multimodal framework for video generation that accepts text, reference images, audio, and pose inputs simultaneously. It claims state-of-the-art performance across diverse conditioning settings.

85% relevant

OpenBMB's VoxCPM 2: 2B-Param Open-Source TTS for Multilingual Voice

OpenBMB launched VoxCPM 2, a 2-billion-parameter open-source text-to-speech model. It generates multilingual, emotionally expressive speech from text descriptions and runs on consumer-grade hardware.

97% relevant

Automate Your MCP Server Marketing with This GitHub Actions Pipeline

A GitHub Actions tool that automatically markets your MCP server or Claude Code skill across platforms and directories, saving hours of manual promotion work.

100% relevant

Alibaba Paper Shows AI Moving Beyond Text, Echoing Pichai's Warnings

Alibaba has published a research paper illustrating AI's progression beyond pure text generation. The work serves as a concrete example of the accelerating, multi-modal capabilities that industry leaders like Google's Sundar Pichai have recently cautioned about.

75% relevant