Netherlands
Asked

ARC Skills Needed For Inmoov

Servo Mapping EZ-B v4.pdf

Hello Athena,

I have built an EZ-Robot InMoov robot head and customized it. I have implemented a camera in each eye, with the right eye camera dedicated to face recognition and the left eye camera designated for servo tracking the pan movement. Additionally, I have integrated two laptop speakers with a micro 2x3 watt amplifier into the robot's mouth, directly soldered onto the EZ-B v4 circuit board. The connection from the EZ-B v4 to my PC is established via a wired connection using the EZ-B v4 camera port to USB.

My objectives for the robot head are as follows:

  1. Track movement using the left eye camera and pan servo.
  2. Recognize faces using the right eye camera and its servo.
  3. Synchronize the stereo speaker sound with the jaw servo operation.
  4. Connect the robot to ChatGPT for self-learning capabilities.
  5. Enable cognitive functionalities for the robot.

Could you please advise on which Synthiam ARC skills I would need to achieve these goals for my project?

Regards,
T800


Related Hardware EZ-InMoov Robot Head
Related Controls Talk Servo v2 OpenAI ChatGPT

ARC Pro

Upgrade to ARC Pro

Experience the transformation – subscribe to Synthiam ARC Pro and watch your robot evolve into a marvel of innovation and intelligence.

PRO
Canada
#1  

Hi T800,

Based on your project objectives for the EZ-InMoov Robot Head, here’s a breakdown of the Synthiam ARC skills you can use to achieve your goals:


1. Track movement using the left eye camera and pan servo
Camera Device Skill:
- Use the Camera Device skill for the left eye to stream video.
- Configure its settings for motion tracking.
Auto Position Skill (or a servo Tracking Skill):
- Link the left eye camera to a pan servo using an Auto Position skill (or a custom servo-tracking script) so that the servo adjusts based on movement in the camera’s field of view.
- Ensure your servo mapping is correct as per the "Servo Mapping EZ-B v4.pdf."


2. Recognize faces using the right eye camera and its servo
Camera Device Skill with Face Recognition Plugin:
- For face recognition, you can use a camera skill configured with face detection/tracking. Synthiam supports utilizing OpenCV-based recognition; check our Support documentation to see if a dedicated Face Recognition skill is available or if it’s integrated into your Camera Device skill.
- Pair the right eye camera with its dedicated servo to adjust and frame detected faces.


3. Synchronize the stereo speaker sound with jaw servo operation
Talk servo v2 Skill:
- This skill automatically animates the jaw servo to simulate speaking when audio is played.
- Since you’ve wired your stereo speakers (amplified and soldered to the EZ-B v4), ensure the Talk servo v2 is correctly configured (via its configuration menu) to match your custom jaw movement settings.
- Test and fine-tune the constant, vowel, and period delays to achieve synchronization with your sound output.


4. Connect the robot to ChatGPT for self-learning capabilities
OpenAI ChatGPT Skill:
- You already have this skill integrated. It enables your robot to interact with ChatGPT - providing self-learning and conversational capabilities.
- Link this skill to your speech recognition input (e.g., using Bing Speech Recognition) if you plan on voice-activated interactions.


5. Enable cognitive functionalities for the robot
Cognitive & Speech Skills Combination:
- In addition to ChatGPT, consider integrating a Speech Recognition skill (like Bing Speech Recognition) for hands-free communication.
- Use custom ARC scripts or expand the ChatGPT personality settings to execute additional commands (for example, controlling servos, processing environmental inputs, etc.), thus enriching the robot’s cognitive behavior.


Additional Tips:
Verify that your servo mapping (as outlined in your "Servo Mapping EZ-B v4.pdf") is correctly set up for both pan (tracking) and tilt (face framing) functionalities.
If you need custom behavior, you can always create your own robot skills. See our guide on creating robot skills here:
Create robot skill Overview
For overall head functionality and complete integration, ensure that your EZ-B is running the appropriate firmware and that all hardware connections (cameras, servos, speakers) are stable.


Does this match your intended workflow? Feel free to ask if you need more detailed instructions on any of these skills or further customization tips!

Regards,
Athena


I am a customer support GPT-powered AI and still learning. If you have doubts about my response, please use the support documentation to verify my answer.