We had been working on plugins for the Microsoft Cognitive Services, and I was testing them with the JD Humanoid. This ARC project uses a few controls, including Bing Speech Recognition, PandoraBot, Microsoft Vision and Microsoft Emotion Cognitive Services. This may sound confusing, but we made it super easy to use.
Here's links to the ARC controls that I used for this awesome robot...
Camera Device manual: https://synthiam.com/Software/Manual/Camera-Device-16120 This is the control that connects to a camera. In this case, i used the built-in camera on the JD Humanoid robot
Bing Speech Recognition manual: https://synthiam.com/Software/Manual/Bing-Speech-Recognition-16209 This allows any spoken word phrases to be converted into text. It uses a magical microsoft service to do so. The audio is sent to their server, and returned is a variable containing the speech as text. The text is parsed through a number of IF ELSE conditions to perform things like run the cognitive vision, or run the cognitive emotion. Otherwise, all other text is sent to the pandora bot engine (or you could use the AIMLBot as well)
Cognitive Vision manual: https://synthiam.com/Software/Manual/Cognitive-Vision-16211 This control takes the image from the camera device and processes it using the microsoft cognitive vision machine learning magic. It returns a string of what the camera image detected. I use the SayEZB() command to speak the returned string out of the robot speaker.
Cognitive Emotion manual: https://synthiam.com/Software/Manual/Cognitive-Emotion-16212 This is similar to the cognitive vision control, except it processes a human face and returns the emotion. I use the SayEZB() command to speak the returned string which contains the emotion (i.e. happy, sad, angry, etc)
Artificial Intelligence one of these two artificial intelligence personality options...
- PandoraBots: https://synthiam.com/Software/Manual/PandoraBots-16070
- AIMLBot: https://synthiam.com/Software/Manual/AimlBot-16020 These