Skip to main content

面部追踪

蓝图自定义

在开播准备环节,您可以点击 面部追踪进阶设置… 来对蓝图进行定制。

Customizing face tracking.

目前我们提供以下几种选择:

  • BlendShape 映射: 选择适合您模型的 BlendShape 映射。比如说,如果您的模型带有 "Perfect Sync"/ARKit Blendshape, 请您选择 Arkit; 如果您的模型是由MMD模型转换而来,请选择 MMD ;其他请选择 VRM。 Warudo 会尝试自动识别模型类型,但您也可以在这里手动更改类型。-
  • 启用头部和身体运动: 此选项开启时,Warudo 会尝试基于动捕数据来同时运动模型的头身部分。在您使用完整的全身动捕系统的情况下,此选项会被开播助手设置为关闭
  • 头部待机动画 (自动眨眼/自动眼部动作/自动摇头): 在开启状态下,Warudo 会给角色增加微小的头部、眼部动作。
  • 视线: 此选项开启时,,您的角色视线会望向指定方向(默认指向摄像机位)。这样可以帮助您在任何机位下保持与观众的眼神交流并允许您的头部自由转动。
  • Lip Sync: If enabled, Warudo will animate your character's mouth based on your speech. You can choose to enable lip sync only when tracking is lost, or always enable it.
信息

These options above affect the generated blueprint; therefore, to change these options after the setup, you need to re-run the setup, or manually modify the face tracking blueprint. Character → Motion Capture → Blueprint Navigation provides shortcuts to the specific parts of the blueprint that you may want to modify:

Tracker Customization

After onboarding, you can go to the corresponding face tracking asset (e.g., iFacialMocap Receiver, MediaPipe Tracker) to customize the tracking data itself. The following options are available:

  • Mirrored Tracking: If enabled, Warudo will mirror the tracking data.
  • BlendShape Sensitivity: Adjust the sensitivity of the character's facial expressions. If your expressions are too subtle, increase the sensitivity; if your expressions are too exaggerated, decrease the sensitivity.
  • Configure BlendShapes Mapping: Adjust each individual blendshape's threshold and sensitivity. This is useful if you want to disable certain blendshapes, or if you want to adjust the sensitivity of each blendshape individually. The left two values are the range of the input blendshape value from the tracker, and the right two values are the range of the output blendshape value to the character. For example, if the input range is 0-1 and the output range is 0-2, then when the input blendshape value is 0.40, the output blendshape value will be 0.80.
    • To disable a blendshape, set the top right value to 0.
    • To make a blendshape more sensitive, increase the top right value; to make a blendshape less sensitive, decrease the top right value.
    • To trigger a blendshape at a higher threshold (e.g., your character mouth is already slightly opened but your mouth is still closed), increase the bottom right value. To trigger a blendshape at a lower threshold, decrease the bottom right value.
  • Head Movement/Rotation Intensity: Adjust the intensity of the head movement/rotation.
  • Body Movement Intensity: Adjust the intensity of the automatic body movement due to head movement.
  • Body Rotation Type: Makes the body rotate naturally when the head rotates. When set to None, the body will not rotate when the head rotates. When set to Normal, the body rotates in the same direction as the head. When set to Inverted, the body rotates in the same direction as the head on the X and Y axes, but in the opposite direction on the Z axis. You can also set Body Rotation Type to Custom, which allows you to customize both the direction and magnitude of each rotation axis.
  • Eye Movement Intensity: Adjust the intensity of the pupil movement.
  • Eye Blink Sensitivity: Adjust the sensitivity of the eye blink; this is a shortcut for adjusting the sensitivity of the eye blink blendshape in Configure BlendShapes Mapping.
  • Linked Eye Blinking: If enabled, force both eyes to blink at the same time. Useful if your tracking is not accurate enough to blink each eye individually.
  • Use Bones/BlendShapes for Eye Movement: Whether to use eye bones or blendshapes for eye movement. If your model has eye bones, it is recommended to use eye bones, as they are more accurate and allow for IK. There are two cases where you may want to enable Use BlendShapes For Eye Movement:
    • Your model does not have eye bones.
    • Your model's eye blendshapes are supplementary to the eye bones, i.e., the eye blendshapes change the shape of the eyes, in addition to the eye bones moving the pupils. (Your modeler should be able to tell you whether this is the case.)

Frequently Asked Questions

When I move my head, my character's body also moves. How do I disable this?

This is caused by the Body Movement Intensity option in the motion capture receiver asset. Set it to 0 to disable body movement.

How do I enable lip sync?

During the onboarding process, you can enable lip sync by clicking Customize Face Tracking... and enabling Lip Sync.

You can adjust lip sync settings after onboarding by clicking Character → Motion Capture → Blueprint Navigation → Lip Sync Settings.

My model's mouth is slightly open even when my mouth is closed.

This is usually caused by applying ARKit tracking on a non-ARKit-compatible model. To fix this, you can either:

  • Add "Perfect Sync" / ARKit blendshapes to your model. This is recommended since ARKit-based face tracking is much more expressive, and you already have the tracking data.
  • Click Configure BlendShapes Mapping in the tracker asset, and increase the threshold (i.e., set the bottom right value to a negative number like -0.25) of the ARKit mouth blendshapes jawOpen, mouthFunnel, mouthPucker.

My body rotates in the opposite direction when I turn my head.

By default, Body Rotation Type is set to Inverted to mimic Live2D models and achieve a more anime look. If you want to disable this, set Body Rotation Type to Normal or None.

翻译者

上次更新于 2024.11.16