⚡ v1.0 — Blender 4.2+

Motion Gen

Type a sentence, get a full 3D character animation. Motion Gen bridges natural language and physics-based motion synthesis directly inside Blender — no mocap studio, no keyframing, no pipeline juggling.

Text-to-Motion Diffusion Model Blender Integration Auto Rigging GPU & CPU

What is Motion Gen?

Motion Gen is a Blender addon that lets you generate realistic 3D humanoid animations from plain-English text prompts. Under the hood it orchestrates a multi-model AI pipeline: a large language model (Qwen 3-8B) interprets your prompt and translates it into a structured motion description, a CLIP vision-language encoder aligns the semantics, and a 1-billion-parameter diffusion network synthesises the actual joint rotations and root translations frame-by-frame.

The result is automatically mapped onto a built-in SMPL-X character armature and imported into your Blender scene — ready to render, retarget, or refine. The entire process runs locally on your machine; no cloud API, no subscription, no internet required after initial setup.

📢
Public Disclosure: This add-on connects to external services (PyPI and Hugging Face) to download required open-source AI weights and Python libraries. Specifically, it requires downloading two large-scale AI models: Tencent HY-Motion 1.0 and Qwen 3-8B.

Users must explicitly agree to the Tencent Community License and the Qwen License via the in-Blender setup panel before these models can be downloaded or used. No data is sent without explicit user input.

What You Get

A complete prompt-to-animation pipeline, entirely offline and deeply integrated into Blender.

Prompt-to-Motion

Describe any human motion in natural language — from "a person casually walking" to "a martial arts spinning kick" — and watch it materialise as keyframed animation data.

Isolated ML Runtime

A fully standalone Python 3.11 environment lives inside the addon folder. Zero conflicts with Blender's built-in Python — install once, forget forever.

🦴

Auto-Rigging & Import

Generated motion is automatically applied to a bundled SMPL-X armature model. The character appears in your scene with all keyframes baked and ready.

🎛

Fine-Grained Controls

Tune inference steps, CFG scale, seed, and duration. Use Draft Mode (20 steps) for rapid iteration, then switch to Production (50-60 steps) for final quality.

🖥

GPU + CPU Fallback

Optimised for NVIDIA GPUs with CUDA, but includes a "Force CPU" toggle for AMD, Intel, and Apple Silicon users — or anyone without a dedicated GPU.

🔒

Fully Offline

After initial model download, everything runs locally. Your prompts and generated data never leave your machine.

Generated Animations

Every animation below was generated entirely from a text prompt — no keyframing, no mocap data.

A person jumping A person jumping
Picking a heavy object Picking a heavy object
Picking a small object Picking a small object
Running and kicking Running & kicking
Sitting down Sitting down
Slow running Slow running

💻 System Requirements

General Software & Storage

  • OS Windows 10 or 11 (64-bit) for the automated 1-click installer. Mac and Linux are supported but require manual terminal setup.
  • Host Software Blender 3.0 or higher.
  • Storage 15 GB – 20 GB of free space (SSD highly recommended).
  • Internet Required for the first run to download the AI models and Python environment.
Tier 2 — Fallback

No GPU / Mac / AMD

Enable the Force CPU toggle. Entire multi-model pipeline runs in System RAM.

  • GPUNone required.
  • RAM32 GB absolute minimum (64 GB highly recommended).
  • CPUHigh-end modern processor strongly recommended.
💡
Why so much RAM? Without VRAM, your computer must hold the OS, Blender, Qwen3, CLIP, and the 1-billion-parameter motion network entirely in System RAM simultaneously. 16 GB machines will crash.
⚠️
Speed Warning: CPU rendering is highly accurate but severely bottlenecked. A 3-second animation that takes ~15 s on GPU will take 10–15 minutes on CPU. Please be patient.

Getting Started

Three steps: install the Blender addon, set up the backend, then import the AI model weights.

Install the Addon in Blender

  • Download the motion_gen.zip file (do not extract it).
  • Open Blender (version 4.2.0 or higher recommended).
  • Go to Edit → Preferences → Add-ons.
  • Click Install… at the top right.
  • Locate and select motion_gen.zip, then click Install Add-on.
  • Enable the addon by checking the box next to Animation: Motion Gen.
Installing addon zip in Blender Preferences

Install the Motion Gen Backend

  • Once enabled, expand the addon details by clicking the arrow next to its name.
  • In the Local Setup section, you will see the status of the Python runtime environment.
  • Click Install Python Runtime (or Install Motion Gen Backend). This will configure a standalone Python instance just for the addon.
  • Wait for the installation to finish. You can check the Blender system console (Window → Toggle System Console) for progress.
  • Wait until the status says Runtime and Dependencies Installed!
Python runtime installation progress

Download & Import AI Models

Due to their large size, the AI models must be downloaded separately.

  • In the addon preferences under AI Model Paths, click the provided links to download:
  • Use the file selectors in the addon preferences to point to the downloaded files.
  • Click Import Selected Models. The addon will copy these files to the correct internal directories.
  • When finished, it will display All required models found locally!
Model download and import process
💡
Tip: After all three steps are done, open the 3D Viewport, press N to open the Sidebar, go to the Motion Gen tab, type a prompt, and click Generate & Load Motion.
Full generation workflow

Understanding the Panel

Every control in the Motion Gen sidebar panel explained — from prompt input to the final generate button.

Motion Prompt

Describe the animation in plain English. Focus on concrete physical actions. Keep under 60 words.

Type: Free text  ·  Default: A person jumping over a low wall

Duration (s)

Animation length in seconds. Longer = more frames but slower generation. Caps at 12s (360 frames).

Range: 0.5 – 30  ·  Default: 3.0

Draft Mode

Locks steps to 20 for fast rough previews. Turn off for final quality.

Type: Toggle  ·  Default: Off
🔢

Steps

Diffusion iterations. More = smoother physics. 20–30 for drafts, 50–60 for production.

Range: 10 – 100  ·  Default: 50
🎚

CFG Scale

Prompt adherence vs. physics. Low (2–4) = natural flow. High (5–7) = strict prompt. Above 8 may jitter.

Range: 1.0 – 10.0  ·  Default: 5.0
🎲

Seed

Set -1 for random variation each time. Use a fixed number to reproduce a specific result.

Range: -1 – ∞  ·  Default: -1

Generate & Load Motion

Runs the full pipeline — encodes your prompt, generates motion via diffusion, and imports a rigged character with the baked animation. Shows elapsed time and live status during generation.

Motion Gen panel interface

The Motion Gen panel lives in the 3D Viewport sidebar. Press N to toggle it.

🦴 Retarget to Your Own Character

Motion Gen generates animation onto a built-in SMPL-X armature. To transfer that animation to your own custom character, use Auto-Rig Pro's Remap tool. Here's a step-by-step walkthrough.

📦
Prerequisite: You need the Auto-Rig Pro addon installed in Blender. The Remap feature is included with the addon. Full official docs: ARP Remap Documentation

Generate Your Animation

Use Motion Gen as usual — type your prompt, adjust settings, and click Generate & Load Motion. This creates a rigged SMPL-X character with baked keyframes in your scene. This will be the Source armature.

Import or Prepare Your Target Character

  • Ensure your custom rigged character (the Target armature) is in the same scene.
  • It can be any type of armature — Auto-Rig Pro, Rigify, or a custom rig.
  • Important: Zero out the armature object's Location and Rotation. Press Alt+G and Alt+R, or manually set them to zero in the N-panel.

Open the Remap Panel

  • Go to the 3D Viewport → Sidebar (N-panel) → ARP Tab → Remap.
  • Select the Motion Gen SMPL-X armature and click the Source field to assign it.
  • Select your custom character armature and click the Target field to assign it.
  • Click Auto-Scale to match the source armature's scale to the target, or manually scale with S.

Build & Configure the Bones List

  • Click Build Bones List to populate the mapping.
  • If your target is an Auto-Rig Pro armature, you can load a built-in preset (Mixamo, Rokoko, etc.) using the dropdown arrow button.
  • Otherwise, manually map each source bone to its corresponding target bone. Tedious but straightforward — match hips→hips, spine→spine, etc.
  • Mark the root bone (typically hips/pelvis) with Set as Root — this enables translation + rotation retargeting for that bone.

Match Rest Poses

  • Both armatures must face the same direction in rest pose. Enter Edit Mode (Tab) on each to verify.
  • If the rest poses differ (e.g. A-pose vs T-pose), click Redefine Rest Pose on the source armature.
  • Tick the Preserve checkbox to keep the armature untouched while saving a new internal rest pose.
  • Select mismatched bones (arms, legs) and click Copy Selected Bones Rotation to auto-align them to the target's orientation.
  • Click Apply when done.

Retarget!

  • Click Re-Target to transfer the animation.
  • If proportions differ significantly, use Interactive Tweaks to adjust bone positions and rotations after retargeting.
  • For example: if the pelvis is too high, select it in the bones list, set an Additive Location value, and click the appropriate axis button.
💡
Tip: Once you've created a working bone mapping for your character, export it as a preset using the export button in the Remap panel. This way you can instantly retarget future Motion Gen animations without re-mapping every time.
⚠️
Note: Interactive Tweaks currently only work with Euler rotation mode. If your target rig uses Quaternions, select all bones in Pose Mode and Alt+Click the rotation mode to switch all bones to XYZ Euler before tweaking.

Prompting Guide & Best Practices

Getting great results starts with writing great prompts. Follow these guidelines to get the most out of Motion Gen's text-to-motion pipeline.

🌐

Language & Length

Please use English. For optimal results, keep your prompt under 60 words.

🎯

Content Focus

Focus on action descriptions or detailed movements of the limbs and torso. Concrete physical actions produce the best results.

Current Limitations NOT Supported

Non-humanoid Characters

Animations for animals or non-human creatures.

Subjective / Visual Attributes

Descriptions of complex emotions, clothing, or physical appearance.

Environment & Camera

Descriptions of objects, scenes, or camera angles.

Multi-person Interactions

Motions involving two or more people.

Special Modes

Seamless loop or in-place animations.

Example Prompts

01

A person performs a squat, then pushes a barbell overhead using the power from standing up.

02

A person climbs upward, moving up the slope.

03

A person stands up from the chair, then stretches their arms.

04

A person walks unsteadily, then slowly sits down.

Known Limitations

Motion Gen is powerful, but understanding its boundaries will help you get the best results.

Addon Limitations

Currently supports a single built-in SMPL-X character template. Custom character retargeting is not yet available. The addon requires a standalone Python 3.11 runtime (~3 GB) in addition to the model weights.

Model Limitations

The diffusion model generates humanoid motion only — no quadrupeds, no multi-character interaction, no finger/facial animation. Very long prompts or unusual descriptions may produce unpredictable results. Physics accuracy degrades past ~10 seconds of continuous motion.

Hardware Constraints

NVIDIA GPUs with at least 8 GB VRAM are required for real-time generation. CPU-only execution works but is 40–60× slower. AMD and Intel GPUs are not directly supported for accelerated inference.

Prompt Sensitivity

Results are highly dependent on prompt wording. Simple, descriptive sentences work best. The model responds better to concrete actions ("a person jumps over a box") than abstract concepts ("feeling happy").

Platform Support

The automated 1-click installer only works on Windows 10/11 (64-bit). macOS and Linux users must manually set up the Python environment via terminal.

Generation Determinism

Using a fixed seed produces consistent results on the same hardware. Different GPU architectures or CPU vs GPU may yield slightly different outputs for the same seed.

Changelog

Version Date Highlights
v1.0 April 2026 Initial release — prompt-to-motion generation, SMPL-X auto-rigging, GPU + CPU fallback, manual model import, isolated Python 3.11 runtime, Draft & Production quality modes.

Model & Dependency Licenses

Motion Gen orchestrates several third-party AI models. Use of this addon requires explicit agreement to the following licenses via the in-Blender "Click-Wrap" agreement.

HY-Motion 1.0
Tencent — Motion diffusion network (1B params)
Tencent Community License
Qwen 3-8B GGUF
Alibaba / Unsloth — Language model for prompt translation
Qwen License
CLIP ViT-L/14
OpenAI — Vision-language encoder for semantic alignment
MIT
PyTorch
Meta — Deep learning framework
BSD-3
llama-cpp-python
Andrei Betlen — GGUF inference binding
MIT
⚖️
Legal Disclaimer: The Motion Gen addon acts as a downloader and local interface for third-party AI models. The developers do not distribute the model weights directly. By using the automated setup script, you are entering into a direct legal agreement with the respective model creators (Tencent, Alibaba/Qwen). The "Click-Wrap" agreement inside the Blender UI must be accepted before any third-party software is downloaded or executed.