The first working prototype was Unity writing images to disk, and having a python poll the disk for a file. Slow, 4-5 per second. Using a socket, the bottleneck becomes processing time in the python server file classify_server.py. Each classification takes .8 seconds to load the model each time. With the model loaded once at first, classification takes .175 seconds. The iOS version from the tensorflow for IOS webpage, seems to be much faster on an iPhone X. But iOS requires a "flat", stripped-down model, which is probably faster already.