As a quicker solution I've made this modified version of AutoLipSync that uses a blend shape instead of a jaw bone to animate based on the audio volume.
This seems like a good quick way of doing basic lip syncing with Mixamo characters.
https://gist…
Updated link to relevant Mixamo thread
Could you provide some info on how you'd approach writing a custom script for this? Essentially each phoneme would need to be mapped to a set of values for Mixamo's dozens of blendshapes. Like
Mouth Left: 5.1Mo…