Search this site
Embedded Files
Seungjun Lee
  • Home
  • Non-ML Projects
Seungjun Lee
  • Home
  • Non-ML Projects
  • More
    • Home
    • Non-ML Projects

Seungjun Lee

lsj3285007@gmail.com

X(formely Twitter) GitHub Linkedin

Experience

Google Summer of Code 2023 @ TensorFlow Archive

  • Made a custom text summary dataset with 12,000 examples encompassing 13 types of document and average length of summary/document is 19% using PaLM API making it superior to publicly available dataset such as Xsum or CNN-DailyMail in Summary length and content diversity

  • Using this created text summary ml models fine-tuning T5, and GPT2. Doing experiments found out that fine-tuning GPT2 setting loss function to give same importance to original document and summary part yields the best result in terms of Rouge-L score.

ML Projects

Toxic - LLama GitHub

  • Engineered a method to jailbreak the LLaMA-3.2-1B-Instruct model using PPO, successfully reducing the refusal rate on 131 unsafe prompts from 91.6%  6.1%.

  • Innovated a training approach that exclusively used a safe prompt dataset and an RLHF dataset, bypassing the conventional need for SFT with toxic prompt-response pairs.

  • Evaluated model performance trade-offs, showing that while overall IFEval benchmark scores slightly decreased, the model achieved significant improvements in targeted instruction-following, including a +16% gain in case handling and +10% in constrained response generation.

Llama-finetuning GitHub

  • Developed a modular framework for fine-tuning Llama models on any text dataset, implementing LoRA and DoRA from scratch.

  • Fine-tuned the Llama 3.2-1B model on SQuAD v1, achieving cross-entropy losses of 1.5245 and 1.3891 on the SQuAD v1.0 dataset for LoRA and DoRA, respectively.

Image Colorization GitHub

  • Developed a flexible framework for training various generator architectures (U-Net, ResNet-backbone U-Net, and Attention U-Net) based on the Pix2Pix architecture for image colorization.

  • Trained on the COCO2017 dataset, achieving LPIPS scores of 0.2150, 0.2098, and 0.2247 for the respective models.

Pic Scribe GitHub

  • Developed and implemented an image captioning model using a Vision Transformer (ViT) encoder and a Transformer decoder architecture.

  • Trained the model on the COCO2017 dataset, achieving 0.640 CIDEr score on COCO2017 validation set.

  • Successfully deployed the model as an iOS application, leveraging CoreML for efficient on-device inference and SwiftUI for a seamless user interface. The app allows users to easily upload images and receive descriptive captions.

ML Implementations GitHub                                                                                        

  • Implemented Linear, Convolution, TransposeConvolution, MaxPool, RNN, LSTM, Embedding, MultiHeadAttention, Transformer, K-means clustering, and KV Cache using NumPy.

  • Implemented ViT, DenseNet-BC, DDColor, ResNet, YOLO, CRNN, FlashAttention, Diffusion (DDPM), ModernBERT, Conformer, MQA, GQA, and MLA using PyTorch.

Animal Classifier GitHub                                                                                        

  • Develop a web application that classifies a provided animal photograph into one of ten categories by creating and deploying a Flask app to Google Cloud Run.

  • Created a model using InceptionV3 as base, and trained it dataset from Kaggle, and achieved an 85% of accuracy on validation set.

Checkout my Non-ML Projects

Awards

Apple Swift Student Challenge 2021

  • Selected as one of 350 winners by building a net-worth tracking app that shows an allocation of net-worth in two hierarchies with pie charts.

MLH hack This Fall

  • My team and I took the 1st place among 4,000 students and awarded with 50,000 INR and $2000 USD worth of gift cards by creating an iOS app that helps people with speaking disorders to communicate by transforming emoji/text to audio using AVFoundation framework.

Google Sites
Report abuse
Page details
Page updated
Google Sites
Report abuse