- python>=3.8
- mediapipe
- pytorch (cpu version is ok)
- blender >= 3.0 (for reading assets and binding animation)
-
Prepare your character model
Currently this script uses Blender to load model skeleton and bind animation. Your model should be saved as .blend file.
You may edit your model to assure that
-
Model must be in rest pose (clear all bone rotation/translation/scale in pose mode). And the rest pose should be close to T pose.
-
Clear previous bone animation data and constraints.
-
Name related bones as below (in lower case). You may refer the mixamo example assets/mixamo.blend to see to name the bones so that they can be recogonzed and binded.
-
Save the model as
.blend
file somewhere.
-
-
Run the script
mocap.py
.python mocap.py --blend your_character_model.blend --video your_video.mp4 [other options]
The program will read and capture motion from the video, save the animation data, and then open Blender and bind the animation to your charactor model. After everything is done, you should be able to see the Blender window with your character already animated.
- Now working on face capture.