My robot would be sitting at a table, and it would look down at the objects on that table. It would be able identify the objects on the table: salt, pepper, a can of beer, a knife, a plate a spoon, and carrots on the plate. It would also know the coordinates of each object in 3D space of each object. Knowing that it would reach out and pick up the beer. I think Tensorflow can do some of this already. Microsoft Cognitive vision gives you the objects it sees, but not the locations.
Other robots from Synthiam community

Jayutmb's My Ez Robot R2 Project :)
So here is a dump of my ez robot R2 project he is not near done yet but i have been pretty lazy in posting updates to...

Bhouston's My Inmoov Running On Synthiam
Initial test of my InMoov robot running on the new EZBv4. I have 24 servos connected to one v4 and all are running OK....

Jstarne1's Ladybug Lawn Mower 3D Designing The First 3D...
Hello All, I am sharing early images of one of my current projects that is actually meant to be released as a kit. This...