Future Directions

The first step to take in terms of future directions is to accommodate for the issues we encountered during our development and training of our model, specifically the limitations in our dataset, our masks, and the latter inpainting architecture.

  • With regards to our dataset limitations, some approaches we could take are to collect our own data in the field, augment our current dataset through translations and cropping, or otherwise artificially apply raindrop distortions and blurs to images of scenes. This would allow us to increase the size of our dataset and thus provide our models with more examples to learn from.

  • In order to improve the quality of our raindrop masks, we would consider approaching the problem through less traditional image transformations (i.e. morphology operations, filters, and Canny edge detector) and instead take the time to label masks from the rainy image. This would allow for the training of convolutional neural networks through supervised learning to learn how to automatically generate the raindrop masks through semantic segmentation of the image into raindrop and non raindrop pixels.

  • We also consider changing the inpainting architecture into a conditional GAN. Due to the fact that we completely remove portions of the image in the process of masking out the raindrops, oftentimes the GAN does not have enough information in order to properly reconstruct a realistic representation of the derained scene. Implementing a conditional GAN would allow the model to retain a correspondence between the input and output and thus learn to derain scenes more realistically.

In terms of future developments past the corrections, our model is currently only trained to handle raindrops of varying sizes on windows and camera lenses, creating distinct regions of distortion to detect and replace. However, there are a multitude of other rain scenes that may be encountered outdoors. For example, there could be heavy rain creating large rain streaks across the image, the rain could be striking different surfaces and creating rain splatters in the image, or the rain could creating a mist and fog up the scene. Further developments in our model would seek to take into account these varied rain scenes in order to make our model applicable to any type of rain that it may encounter, thus resulting in an all purpose deraining model.