Our project implements an adversarial example on Instagram's OCR algorithm. Since we are not aware of the exact model, we proceeded with a Black Box approach to address the problem.
The implementation was done in phases wherein we performed a targeted attack with antonyms and then proceeded to untargetted attacks that covered flagged content on Instagram such as Vaccinations and COVID-19. More details for our implementation can be found here.
A target image along with an input image is given to alter the output of the OCR model to the text present in target-image.
For a given input image containing a single word, a corresponding target image is given containing its antonym.
Images containing terms related to "Vaccination" are taken. We manually modify the images hiding the word "Vaccination" to create target-images.
No target image is given. The goal is just to alter the OCR model output for the given input-image.
No target image is given. The goal is to just alter the model output.
[Success Rate = No. of perturbed images able to fool the model/Total no. of Images considered]
Upon manually uploading statuses on Instagram, these are the success rates that we were able to achieve on the platform