Visual explainability is a subset within neural network interpretability research, specifically geared towards explaining neural network decisions to users. The large deep learning algorithms are often seen as 'black-box' models where the decisions are not understood. The goal of an explainability algorithm is to investigate the reasons behind 'Why?' a particular decision is made. However such 'why?' questions are hard to answer uniformly. Instead, we advocate for a contrastive approach where explanations must answer 'Why P, rather than Q?' questions...
Neural networks reason inductively - after having learnt necessary and sufficient patterns while training, they search for similar patterns at inference. Having discovered these patterns, they make their decisions on given data. According to the network, the patterns are the 'cause' that lead to the 'effect', in this case the decision. However, when the train and test distributions don't match, networks fail to make the right decisions. Research in the distribution shift generally falls under robustness. We advocate for an abductive reasoning approach - create a hypothesis and tests its validity without considering the cause. We propose Introspective Learning...
The science of uncertainty quantification (UQ) deals with assigning probabilities regarding decisions made under some unknown states of the system. Generally, unknown states can occur due to: (1) Lack of data, (2) Underspecified or incorrectly chosen models or test data distributions, (3) Noisy ground truth labels, or (4) Interventions during inference. By definition, any measurement of a quantity requires 'fixing' (or holding constant or intervening) some other quantity for systematic study. If such 'fixings' are conducted across all possible sets of interventions, then no uncertainty exists. Very rarely are all interventions feasible (or possible)...
Trust is an esoteric quantity that cannot easily be measured. To garner the trust of humans, the underlying ML models must lend themselves to certain attributes that fall under the umbrella terminology of trustworthiness. However, these attributes are functions of the scientific communities that consider them. For instance, in the field of autonomous vehicles, the algorithmic trust in the vehicle’s perception module is different from the moral trust that is placed upon it, which is again different from the governmental policy trust that the vehicle abides by. Prediction trust, specifically evaluates trust that can be placed in a particular decision. Uncertainty and trust differ from each other...
PointPrompt is the first visual segmentation prompting dataset based on the Segment Anything Model (SAM). It is a comprehensive collection of human-generated prompts across 6000 images corresponding to 16 image categories and 4 data modalities (natural, seismic, medical and underwater). The prompting data was generated in an interactive manner, and at each step of the prompting process, the generated mask and associated score were shown to the annotator so they could adapt their strategy or move on to the next image. We have compared the segmentation scores obtained by our 48 human annotators against several existing automated prompting methods, showing that human prompting is consistently superior...