I'm a fairly advanced user, but I've come across something that I just can't wrap my head around. I'm hoping the geniuses here can help.
I'd like to make a number of different mouth position comps in AE and use an expression (or plugin, if it exists) to select each mouth instance as dictated by the comp's audio. For example, I have a rounded mouth comp and I'd like it to appear when the speaking character makes an "oh" sound. I have a close-mouthed, teeth-clenched comp that appears when the character makes a "t" sound. Etc...
Basically I'd like to automate my character speaking.
I can easily make the mouth go up and down to the audio, but I'm looking to do something more advanced; I'd like to have it choose between several mouth positions so it looks more realistic. They don't need to morph between mouths; I just want them to appear, as in proper animation.
I was hoping to find a way of having the software listen to the audio and choose the correct mouth for me. I've tried it with Soundkeys with mixed results. I was wondering if someone had another way beyond Aharon's method of doing it by hand.
I did read the post you directed me to. And, as it didn't answer my question, I created this thread.