My robot would be sitting at a table, and it would look down at the objects on that table. It would be able identify the objects on the table: salt, pepper, a can of beer, a knife, a plate a spoon, and carrots on the plate. It would also know the coordinates of each object in 3D space of each object. Knowing that it would reach out and pick up the beer. I think Tensorflow can do some of this already. Microsoft Cognitive vision gives you the objects it sees, but not the locations.
Other robots from Synthiam community

Rb550f's Meet Ewa
This is my 16th robot project EWA. Built from a Walle Ucommand toy. It has a 3d printed head from Xlrobots. I added a...

Justinratliff's Walle Built With Iotiny
Im not brave enough for live hacks yet, but this shows me updating my software, explaining how I built him, trying to...

Mcjeff0125's Robie Sr, Dr. Theopolis Hybrid
May I present to you my Robie Sr, EZ-Robot mod. As you can see, I did not use the stock Robie head. My main reason was...