Asked
— Edited
The IBM Watson Services plugin created by @PTP currently allows you to perform Speech to Text and Text to Speech as well as visual recognition using Watson Services.
You can download and install the plugin here https://www.ez-robot.com/EZ-Builder/Plugins/view/251
Before you can use the plugin, you will need to apply for a free 30 day trial of IBM Speech to Text Account here https://www.ibm.com/watson/services/speech-to-text and a free 30 day trial of IBM Text to Speech https://www.ibm.com/watson/services/text-to-speech/
I will create some examples as it moves forward and try to answer any how to questions.
Thanks for creating this PTP, as with all your plugins this is an excellent piece of work and showcase of your talents.
I spent about 10 hours of photographing, uploading, training (Takes about an hour to train now with over 2000 photo's) using an EZ-B and Camera placing cards on table and changing position / lighting. I had about 80% of the deck with very high accuracy. If it didn't read a card correctly I would take more photos, upload train and repeat.
It can take anywhere from 2 to 20 seconds to come back with an answer now when you show a card.
Now confident I was ready, I pulled out JD and held up a card in front of him. Not a clue (about 5% accuracy).
This tells me my training method is bad or we need a very controlled environment.
I think we could probably do a controlled read if we used a static base with controlled lighting, Example a Game of Blackjack (21) between JD and Six and me as the dealer, with JD and Six sitting in specific positions and cards placed on playing board in predetermined locations.
The other alternative is I create a method that is automated for reading cards and taking photo's. Perhaps it is something like hand JD a Card and then have a script run so he can then turn around in a circle and change the angle of the card taking pictures so the lighting, angle and background keep changing. Then when he is finished (Say 100 photos) upload the photos he took and ask for another card.
Thinking ....
Edit - Turns Out JD dexterity is not good enough to read a card he is holding (Can't get good enough angle / distance) so it will require a custom bot to read cards (More Servo's and perhaps a 360 Servo). Also it appear not just the same model camera is good enough, it has to be the exact same camera. If I take photo's with JD and try to read them on EZ-B + Camera doesn't work and vise versa.
@Nick,
I'm still working on it and researching other options. I've compiled tensorflow from source code, took me 3 failed attempts, successful compilation took 4 hours (core I7) and almost 6 GB disk space, to obtain a 64 bits c library (dll and lib).I can't find that brand, maybe you can suggest one available both in amazon.ca and amazon.com. That way we could share pictures.
I took pictures too... i wanted to test two different resolutions: 320x240 and 640x480, unfortunately the high resolution pictures are low resolution, the snapshot control ignored the camera control setting.
I'll add a picture/snapshot functionality to the watson plugin.
Can you share a zip of one your cards ?
Hi ptp. Here is one card. I am apprehensive sending photos of all the cards due to copyright laws. Maybe we should Print off some creative commons cards.
NineofHearts.zip
Plan B
JD a record player, an iPad playing a Netflix movie.
Have you seen the .net bindings for tensor flow for a plugin? https://www.nuget.org/packages/TensorFlowSharp/1.6.0-pre1
@DJ,
Yes, the wrapper works on top tensorflow.dll (libtensorflow.dll) available from the tensorflow project.
https://github.com/tensorflow/tensorflow/issues/10817
sources: https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-cpu-windows-x86_64-1.2.0.zip
http://ci.tensorflow.org/view/Nightly/job/nightly-libtensorflow-windows/lastSuccessfulBuild/artifact/lib_package/libtensorflow-cpu-windows-x86_64.zip
the tensorflow.dll is 64 bits.
@Nink,
Can you explain your plan B
Regarding the playing cards, we don't need to exchange publicly picture files, let's find a common deck available on both amazons.
OK My deck was this one on Amazon.ca not on amazon.com although I pay $1.25 and they want $7 on amazon. I can put a deck in the mail if you want, or if you can provide a link to your cards on amazon.ca I can order a deck. https://www.amazon.ca/gp/offer-listing/B01A61ZQI8/ref=sr_1_1_olp?ie=UTF8&qid=1522329419&sr=8-1&keywords=victoria+playing+cards&condition=new
Plan B was to just take 100 photos of each card (1 photo every second) in camera app and have background and angle changing using Ipad and record player then move robot head up and down as I do it while creating shadows and changing lighting. Logic was I could get through this in about 2 hours.
I have been thinking about a good use case for the ability to read cards. Obviously a QR code can be read easily so this makes me wonder if this is a good application for vision recognition, but at least we are learning and will make a good demo.