daanelson/motion_diffusion_model 📝❓🔢 → ❓
About
A diffusion model for generating human motion video from a text prompt

Example Output
Prompt:
"the person walked forward and is picking up his toolbox."
Output
Performance Metrics
145.38s
Prediction Time
323.59s
Total Time
All Input Parameters
{ "prompt": "the person walked forward and is picking up his toolbox.", "num_repetitions": 3 }
Input Parameters
- prompt
- output_format
- Choose the format of the output, either an animation or a json file of the animation data. The json format is: {"thetas": [...], "root_translation": [...], "joint_map": [...]}, where "thetas" is an [nframes x njoints x 3] array of joint rotations in degrees, "root_translation" is an [nframes x 3] array of (X, Y, Z) positions of the root, and "joint_map" is a list mapping the SMPL joint index to the corresponding HumanIK joint name
- num_repetitions
- How many
Output Schema
Output
Example Execution Logs
### Sampling [repetitions #0] 1 2 created 1 samples ### Sampling [repetitions #1] 1 2 created 2 samples ### Sampling [repetitions #2] 1 2 created 3 samples ["the person walked forward and is picking up his toolbox." (00) | Rep #00 | -> sample00_rep00.mp4] ["the person walked forward and is picking up his toolbox." (00) | Rep #01 | -> sample00_rep01.mp4] ["the person walked forward and is picking up his toolbox." (00) | Rep #02 | -> sample00_rep02.mp4] [ "the person walked forward and is picking up his toolbox." (00) | all repetitions | -> sample00.mp4] [samples 00 to 00 | all repetitions | -> samples_00_to_00.mp4]
Version Details
- Version ID
3e2218c061c18b2a7388dd91b6677b6515529d4db4d719a6513a23522d23cfa7
- Version Created
- February 10, 2023