Hi! I'm rather new to EZ-Builder, but a friend and I are trying to put together an animatronic head (from Robotics Squared) that will answer questions and look around, blink, move its mouth, etc.
Right now we're using a slightly modified script that @Rich posted to take the variable from the Sound servo control and open the mouth to different levels.
The sound coming into the Sound servo control is being supplied by SayEZB commands in the Speech Recognition control, so that we can have it answer questions and phrases.
Our problem is that the information going to the Sound servo variable is going a lot faster than the actual speech coming out of the speakers. We're thinking we can fix this by piping raw audio out of one port and into another, reading the levels there as the robot is talking and have the actual audio move the mouth.
Our problem with that solution is that we don't know how to pipe the audio out one of the ports. It doesn't even have to be true audio, just approximate volume levels that are fairly true to the actual speech. Anyone know how to pipe synthesized speech out of one of the connectors?
Any help would be greatly appreciated!
(P.S. I'll try to check this every evening, but we're only meeting once a week to work on the project.)