Developed a large language model for Graph Protocol Uniswap v3, implementing optimized GraphQL queries and the llama-index library to efficiently manage extensive texts and facilitate integration with OpenAI's API.
Conducted governance research with uniswap and a16z through collegeDAO
Conducted on chain analysis and developed 10 Dune or flipside dashboards of shock events spanning across DAOs in the NFT and DeFi space like Nouns, Uniswap, Pudgy penguins, Farcaster and more.
Won ETHOxford hackathon bounty by developing an L2 using Polygon zkEVM and fork uniswap v2 and ran transactions through the DEX to stress test the L2.
Conducted textural analysis on proposals spanning across shock events in Decentralised Autonomous organizations (DAO) across the Non Fungible Token and decentralized finance space such as Nouns and Uniswap
Identified metrics accessing significance of proposals affecting acceptance in DAOs as input for Large Language Model
Built a large language model prototype for graph's uniswap V3 to get students to learn how to build subgraphs and substreams.
Leading a research project that analysis web2 to web3 brands growth, starting with scraping on chain and off chain data from niftygateway to conduct analysis on their growth
Deep Reinforcement learning model for stock trading
Used OpenAPI gym to build a long-term investing model that outperformed S&P 500
Causal Impact of Healthcare Intervention on Spending and Coverage
Conducted a rigorous causal inference analysis using regression models in R to evaluate the impact of healthcare interventions on spending and coverage, successfully distinguished between treatment and control groups and quantified policy effectiveness.
Project Caterpillar: Developed models for freight and inventory optimization involving complex logistics.
Worked on data consolidation from 150+ ERP systems, reducing costs and improving efficiency in parts procurement and distribution, learning Genetic Algorithm using Python (Pandas) and ETL techniques for data manipulation and preprocessing.
Project Accident Occurrence predictions with CSFS (Cost Effective Fisher Score)
Extract real time traffic data for data cleaning and manipulation with Python. Used feature selection methods like Cost-sensitive SVM Classifier and my own CSFS to find key variables to solve the problem of cost imbalance and noisy data feature selection.
SEC Filing Similarity Indexing for Trading Signals
Coded in Wolfram Mathematica to analyze and compare the language and word usage in SEC filings from consecutive years for large-cap technology companies. Investigated the correlation between the similarity of language used in these filings and the companies' stock performance, independent of market returns.
Overview of the effects of Uniswap’s fee structure change implemented on October 16, 2023. Read more here.
This dashboards showcases the growth of Farcaster after launching Frames. Reach more here.
Did analysis on engagement and token pricing of pudgy penguin before and after their pudgy plushie walmart launch. Read more here.
Did analysis on growth and engagement of Nouns before and after their first fork. Read more on this here.
Read more on this here .
University of Illinois at Urbana Champaign
Conducted textural analysis on proposals spanning across shock events in Decentralised Autonomous organizations (DAO) across the Non Fungible Token and decentralized finance space such as Nouns and Uniswap
Conducted on chain analysis and built Dune or flipside dashboards of shock events spanning across Nouns, Uniswap, Pudgy penguins, and more
Identified metrics accessing significance of proposals affecting acceptance in DAOs as input for Large Language Model
Built a large language model prototype for graph's uniswap V3 to get students to learn how to build subgraphs and substreams.
Leading a research project that analysis web2 to web3 brands growth, starting with scraping on chain and off chain data from niftygateway to conduct analysis on their growth
Caterpillar
Developing models for freight and inventory optimization involving complex logistics.
Worked on data consolidation from 150+ ERP systems, reducing costs and improving efficiency in parts procurement and distribution, learning Genetic Algorithm using Python (Pandas) and ETL techniques for data manipulation and preprocessing.
PoweredByPlant - Sustainable Vegan Marketplace
Led a team of 6 to pioneer a successful sustainable luxury subscription startup, achieving 10% year-over-year growth in revenue and 20% customer retention rate with vegan influencer using data driven strategies.
Implemented SQL-based automated reporting for product feedback, reducing churn rate by 15% from customers.
Implemented A/B experiments using Google Optimize increasing conversion rate by 9%- website and 20%- ads.
Optimized SEO and design roadmap for website development using Google Analytics, utilized customer insights to increase traffic by 5% and conversions by 1.9%.
Utilized python for statistical modeling and natural language processing to conduct sentiment analysis that helped prioritize 3 new features with one product reaching 900k in impressions organically and drove customer acquisition.
Analyzed user research data with excel to improve UI UX recommendations, improving click through rate by 10%.
Metcon DNDT
Worked on academic research in AI digital twinning techniques to create virtual replicas of repair engine parts used in brazing process using TensorFlow, CNN and OpenCV, enabling real-time analysis and filed for patent on the process.
Conducted customer interviews and utilized customer insights to research and create product narratives using reverse design engineering process that optimized spec design and resource plan for 3D printed metal engine parts in Marine, Aerospace and Oil and Gas industries.
Achieved a 10% increase in adoption rate through data-driven strategies across multi disciplinary teams like Product, Marketing, and Sales, catering directly to customer preferences and market demands for design wireframe.
NTUC Income
Pitched to build a tier based influencer marketing app for the esports industry after analyzing market trends and competitor landscape, built GTM strategy and MVP, generated ideas and built BRD for mobile app deployment.
Conducted exploratory data analysis to uncover trends, patterns, and anomalies in insurance product usage and customer behavior to drive product usage and adoption for Singapore's first true usage-based insurance
Presented findings in Tableau and recommendations to cross-functional teams, including board members.
PROJECTS
Deep Reinforcement learning model for stock trading
Used OpenAPI gym to build a long-term investing model that outperformed S&P 500 by 5%
Project Accident Occurrence predictions with CSFS (Cost Effective Fisher Score)
Extract real time traffic data for data cleaning and manipulation with Python. Used feature selection methods like Cost-sensitive SVM Classifier and my own CSFS to find key variables to solve the problem of cost imbalance and noisy data feature selection.
SEC Filing Similarity Indexing for Trading Signals
Coded in Wolfram Mathematica to analyze and compare the language and word usage in SEC filings from consecutive years for large-cap technology companies. Investigated the correlation between the similarity of language used in these filings and the companies' stock performance, independent of market returns.
Traffic accidents are a huge social cost that needs to be curbed and prediction of traffic accidents helps officials in implementing strategies for safer roads. Most of the research in this field focuses on using powerful classifiers for higher performance in making traffic accident predictions. This report works on building a high performing model by introducing cost-sensitive feature selection methods as a pre processing step that helps to curb the class imbalance and noisy features problem most accident prediction models face when only a classifier without a pre-processing step is used.
This study introduces three cost-sensitive feature selection methods that factor in imbalanced data and helps in figuring out important features using various performance metrics. The selected features are then used in the Cost-sensitive Support Vector Machine Classifier to check for the performance. K-Nearest Neighbor classifier was used as well to draw comparisons on the performance of the three feature selection methods. Overall, the report has found feature selection to be an effective pre-processing step for traffic accident prediction and of all features used, weather was found to be the most important feature with the highest feature score and AUC score, followed by features such as location, time and speed. The results were comparable to previous researches in determining important features for traffic accident prediction.
Efficient freight and inventory optimization lie at the heart of modern supply chain management, offering organizations the promise of reduced costs and enhanced operational efficiency. This research paper delves into the intricacies of "Project Caterpillar," a transformative initiative focused on developing advanced models for freight and inventory optimization.
In the backdrop of complex logistics, this project undertook the monumental task of consolidating data from over 150 Enterprise Resource Planning (ERP) systems. The objective was clear: to streamline the procurement and distribution of parts, thereby optimizing resource allocation and driving cost reductions.
At the core of this endeavor is the implementation of Genetic Algorithm (GA), a cutting-edge optimization technique. Leveraging Python (Pandas) and Extract, Transform, Load (ETL) techniques for data manipulation and preprocessing, the project meticulously crafted an intricate framework. This framework was designed to adapt and thrive in the dynamic world of logistics, where real-time adjustments and intelligent decision-making are paramount.
More details on the results will be provided soon.