You signed in with another tab or window.Reload to refresh your session.You signed out in another tab or window.Reload to refresh your session.You switched accounts on another tab or window.Reload to refresh your session.Dismiss alert
🔬 Research Intern — Language and Speech Lab, NTU Singapore
Oct 2024 – Present Exposure: LLM Fine-Tuning, PEFT, LoRA, LLaMA Adapters, Emotion Extraction, Prompt Tuning • Researched low-resource LLM fine-tuning methods and emotion extraction from text. • Working on synthetic interviews for depression detection using the DAIC-WOZ dataset, fine-tuned on Reddit data and integrated with emotion detection.
🧠 Research Intern — Visual Computing and Data Analytics Lab, IIT BHU
Aug 2023 – Apr 2024 Exposure: GANs, Graph Neural Networks, Fuzzy Logic, DeepWalk, Node2Vec, Struc2Vec • Re-designed GraphGAN with Wasserstein Loss, improving accuracy from *84.7% → 88.57%. • Implemented DeepWalk, Node2Vec, and Struc2Vec for node embeddings. • Developed a Fuzzy Pre-processing Layer using a modified K-Means algorithm to boost accuracy to 88.95%.
Sep 2024 – Present Exposure: Super Resolution, Quantization, Knowledge Distillation, LipSync, Mixed Precision Training • Worked on inference optimization with Quantization, Knowledge Distillation, and Batch Inference. • Applied Mixed Precision Training and Post-training Quantization on CodeFormer. • Reduced inference time by 25%, and total forward pass time by 50%.
🧪 IBM Research Intern [AI 4 Code Team]
May 2025 – Present
Contributing to AI for Code tooling and research problems.
Exploring techniques for intelligent code understanding and generation.
🧠 Publications
Enriching Pre-Training Using Fuzzy Logic
Published and accepted at an IEEE Conference
Focused on enhancing language representation through fuzzy logic integration into the pre-training phase.
An implementation of the GPT(generative pretrained transformer) model, from scratch, which produces Shakespearean text by training on the dialogues written by Shakespeare along with the GPT Encoder.
An autoregressive forecasting implementation of a LSTM network, NBEATS architecture, ARIMA and SARIMAX regressions, and Autoformer architecture on rupee dollar exchange rates using pytorch, pytorch…