Skip to main content


Whether you are streaming at home or at a professional mocap studio, Warudo has you covered. Currently, Warudo supports the following motion capture systems:

Warudo Pro also supports the following motion capture systems:

What motion capture systems should I use?

If you are completely new to 3D VTubing, we recommend starting with MediaPipe for both face tracking and hand tracking. MediaPipe is a motion capture system built into Warudo, and you only need a webcam to use it. If you happen to have an iPhone, we recommend using the iFacialMocap app for face tracking instead, as it offers much higher accuracy than MediaPipe, and offloading face tracking to the iPhone also reduces the load on your computer.

If you are looking to improve your tracking quality, here are a few suggestions:

  • Use an iPhone for face tracking (via the iFacialMocap app). This is also the best face tracking solution you can ever get. If you don't have an iPhone, you may consider looking into a used iPhone. As long as it is compatible with Face ID, it should work with iFacialMocap.

For best tracking quality, we recommend using an iPhone 12 or newer (iPhone mini also works). Older iPhones may have lower tracking quality.


There are two ways to set up motion capture in Warudo. The first and the preferred way is to use the Onboarding Assistant asset that comes with every scene. Simply click Basic Setup → Get Started to start the onboarding process and follow the instructions.

Setting up motion capture in the onboarding assistant asset.

After the onboarding process is complete, you can select the relevant motion capture assets to further customize your tracking. For example, if you are using iFacialMocap, you may notice Warudo adds some movement to your character's body when your head moves. If this is not desirable, you can set Body Movement Intensity to 0 in the iFacialMocap Receiver asset.

Adjusting the body movement intensity of iFacialMocap.

The second way to set up motion capture is to use Character → Setup Motion Capture. This allows you to set up face tracking and/or pose tracking directly. However, some checking steps in the onboarding process are also skipped.

Setting up motion capture in the character asset.

This method is useful if you need to set up multiple face tracking or pose tracking systems, since the onboarding assistant always remove the existing motion capture setup when you start a new onboarding process.

Whichever method you choose, corresponding tracking blueprints will be generated and added to your scene. For example, if you have chosen iFacialMocap for face tracking and MediaPipe for pose tracking, you will be able to see two blueprints generated: Face Tracking - iFacialMocap and Pose Tracking - MediaPipe. You can modify these blueprints to customize your tracking.


You should almost never create a motion capture asset manually, i.e., by clicking on Add Asset and selecting a motion capture asset, such as iFacialMocap Receiver. This is because the motion capture asset alone only receives or provides the tracking data, but does not connect the tracking data to your character, which needs to be done by blueprints. The onboarding assistant and the Setup Motion Capture function automatically create the necessary blueprints for you.

Frequently Asked Questions

My character is not moving.

If you are using a motion capture system that requires an external application, such as iFacialMocap, make sure the application is running and the tracking data is being sent to Warudo. Also, make sure your computer's firewall is not blocking Warudo from receiving the tracking data; you may need to add Warudo to the whitelist, or temporarily disable the firewall.

You may also want to set your network as a private network, as Windows may block incoming connections from public networks. See this guide for more information.

Some motion capture receivers have a Port option. Make sure the port number matches the port number in the external application.

My tracking is too smooth / too jittery.

Please go to the tracking blueprint (e.g., Face Tracking - iFacialMocap) and locate the Smooth Rotation List / Smooth Position List / Smooth Transform / Smooth BlendShape List nodes. Increase the Smooth Time value to make the tracking smoother, or decrease the value to make the tracking more responsive.


Edit on Github


Last updated on 2024.07.11