How to use Google AI for significant impact for millions of children and women?
How to use Google AI for significant impact for millions of children and women?
This is a open source innovative AI solution as part of
Google Chrome Built-in AI Challenge
Recommended Watch: For comprehensive view of our contribution, request you to checkout this demo video
Our idea promotes creative and responsible use of the power of Google AI for a significant impact on the lives of people who will greatly benefit from these technological advances
Hybrid AI: client and server side While healthcare data is multimodal in nature, it suits to utilize the unique strengths of Multimodal Gemini Models and long context
Offline/Client AI: Given the privacy needs of healthcare usecase, and the child-related photography rights, and the digital divide, the use of offline client AI lends here
We believe in the power of offline AI but that if multimodal support can be enabled in client AI, it would make a lots of sense for helping the world make process on UN Sustainable Development Goals, especially in gender and health and education in the challenges (privacy needs of healthcare usecase, and the child-related photography rights, and the digital divide).
Given large MLLMs cannot be run on the client side, it would make sense for the Google team to consider how to bring multimodal support in client AI. Ideas could be how to use MediaPipe and small LLM with some finetuning to extract visual features from MediaPipe (pose detection or object detection) and then feed those visual features into a text LLM. This can happen on client side. It is possible to fine-tune such a combo of visual features with text features for radical transformation of multimodal support on client AI
Source Code link (click)