Projects

The Systems and Architectures
I design to learn, adapt, and scale.

From retrieval-augmented assistants to full stack ML platforms each project is engineered for performance, reliability, and real world impact.

OpenNoteLM (Inspired by Google NotebookLM)

AI Driven Document Intelligence Workspace

Core Stack
Gemini 2.5 FlashVertex AIVector SearchConvexNext.jsTypeScript+5 others
Impact
Latency: 120-250ms TTFT
Retrieval: <50ms
Zero Idle Infrastructure Cost
System Components
Full RAG Pipeline
Reactive Data Layer
Vector Search Engine
Grounding Framework
Internal System
Full Stack

Engineering Details

OpenNoteLM was built to explore how a lightweight, cost efficient alternative to large document intelligence platforms could be implemented using modern LLM infrastructure. The project focuses on designing a low latency Retrieval Augmented Generation (RAG) system that delivers grounded responses over uploaded PDF collections. It integrates Gemini 2.5 Flash with Vertex AI embeddings, leveraging 768 dimensional vector search for semantic retrieval. The architecture emphasizes retrieval optimization (sub 50ms search latency), structured context chunking to reduce hallucination risk, and fast Time to First Token performance (~120–250ms). Deployed using Vercel for frontend delivery, Convex for backend and database orchestration, and protected via Cloudflare at the edge, the system operates on serverless infrastructure with zero idle compute overhead. From a business perspective, the project demonstrates the ability to design and deploy scalable, cost aware AI systems that balance performance, grounding, and infrastructure efficiency.

Full Technical Stack
Gemini 2.5 FlashVertex AIVector SearchConvexNext.jsTypeScriptPythonServerless Architecturetext-embedding-004Material UIReact
Back to Overview

Portfolio Website with Interactive AI Chat Bot

Apple Liquid Glass Inspired UI/UX

Core Stack
Google Gemini APICustom RAG ArchitectureNext.jsTypeScriptTailwind CSSFramer Motion+5 others
Impact
Production Ready RAG Engine
Adaptive model fallback
Secure environment architecture
System Components
JSON-Based Context Chunking
Keyword Scoring Retrieval
Serverless LLM Integration
Prompt Guardrail System
View System
Full Stack

Engineering Details

The portfolio website was designed as an AI native workspace rather than a static personal showcase, with the intention of demonstrating real world LLM integration in a production style web application. It incorporates a custom RAG layer to dynamically surface structured project context using JSON based chunking and controlled prompt pipelines. The system implements model fallback strategies and exponential backoff handling to maintain resilience under API rate limits. Deployed entirely on Vercel, the application reflects a serverless first approach emphasizing secure API management, prompt guardrails, and optimized inference flow. From a business standpoint, it showcases the practical ability to embed LLM powered features into modern web products while maintaining availability, cost control, and performance awareness.

Full Technical Stack
Google Gemini APICustom RAG ArchitectureNext.jsTypeScriptTailwind CSSFramer MotionClaude Sonnet 4.6Google AntigravityVercelSpotify Web API & OAuthReact
Back to Overview

Interpretation of Metagenomic Data using GenAI

Bridging Metagenomics and Clinical Action through LLMs

Core Stack
DeepSeek-R1 8B LLMPythonFastAPIUMAPSentence-TransformersPubMed-Backed RAG+7 others
Impact
Literature Grounded Reports
Pathogen Ranking: Weighted
Automated Contaminant Filtering
System Components
Signal Processing Pipeline
PubMed-Backed RAG
Pathogen Ranking Engine
Safety-Constrained Reasoning
View System
Full Stack

Engineering Details

This project was motivated by the complexity of interpreting metagenomic next generation sequencing (mNGS) outputs, where microbial signal noise and fragmented literature references can make structured reasoning difficult. The system integrates contaminant filtering and pathogen ranking with a PubMed grounded RAG pipeline to support literature backed analysis. Using DeepSeek R1 for structured reasoning, the workflow transforms raw sequencing outputs into organized, evidence referenced draft reports. The focus of the project is on building interpretable AI pipelines that emphasize source grounding and safety constrained reasoning rather than autonomous diagnosis. From a business and research perspective, it demonstrates how GenAI can assist in organizing complex biomedical data into structured insights while preserving traceability and evidentiary support.

Full Technical Stack
DeepSeek-R1 8B LLMPythonFastAPIUMAPSentence-TransformersPubMed-Backed RAGBioinformaticsNCBI Entrez APIPandasNumPyscikit-learnPostgreSQLOllama
Back to Overview

Superstore Sales Analysis and Visualization

Retail Analytics Dashboard & ETL Pipeline

Core Stack
Power BIPySparkSQLPythonMySQLPandas+1 others
Impact
Processed 50,000+ Transactions
50% Reduction in Manual Reporting
30% Increase in Participation Insights
System Components
ETL Data Pipeline
Interactive KPI Dashboards
Predictive Modeling Layer
Regional Profitability Analysis
View System
Full Stack

Engineering Details

The Superstore Sales project was developed to explore scalable data engineering and analytics workflows for large transactional datasets. Processing 50,000+ retail records using PySpark and MySQL, the system implements a structured ETL pipeline that automates cleaning, transformation, and KPI aggregation. The processed data is centralized in interactive Power BI dashboards highlighting regional performance, profitability trends, and customer retention metrics. The project emphasizes reproducible pipelines and automated reporting over manual spreadsheet analysis. From a business perspective, it demonstrates the ability to translate raw operational data into structured dashboards that enable performance monitoring and data driven decision making.

Full Technical Stack
Power BIPySparkSQLPythonMySQLPandasWeights & Biases
Back to Overview

Pneumonia Detection Using Chest X-Rays

CNN Based Pneumonia Detection

Core Stack
PyTorchCNNDeep LearningPythonMedical ImagingTorchvision
Impact
Accuracy: 79.17% test accuracy
97%+ Training Accuracy
Reduced manual screening time
System Components
CNN Architecture
Data Preprocessing Pipeline
Augmentation Strategy
Evaluation Framework
View System
Full Stack

Engineering Details

This project was undertaken to study the practical challenges of deep learning in medical imaging, particularly model generalization across real world data. A convolutional neural network was trained on 5,800+ pediatric chest X-ray images with data augmentation and dropout regularization techniques applied to mitigate overfitting. The model achieved 97%+ training accuracy and 79.17% test accuracy, highlighting the gap between training performance and generalization in clinical datasets. The primary objective was to understand model behavior in healthcare contexts and evaluate how AI assisted preliminary screening systems could support radiological workflows. From a broader impact perspective, the project demonstrates applied deep learning methodology with careful attention to evaluation rigor.

Full Technical Stack
PyTorchCNNDeep LearningPythonMedical ImagingTorchvision
Back to Overview

Autonomous Data Labeler

Active Learning Data Labeling System

Core Stack
AnthropicPythonActive LearningFastAPIReactPostgreSQL
Impact
80% Reduction in Labeling Cost
Scalable Data Annotation
Hybrid Auto-Human Workflow
System Components
Proposal Engine (Weak LLM)
Human-in-the-Loop UI
Arbitration System (Strong LLM)
Feedback Integration Loop
View System
Full Stack

Engineering Details

The Autonomous Data Labeler was designed to explore cost efficient annotation strategies for scaling machine learning systems. The architecture follows a weak to strong orchestration model in which a lightweight model proposes labels, human reviewers validate uncertain cases, and a supervisory model arbitrates edge scenarios. This layered workflow reduces dependency on fully manual annotation while preserving quality control mechanisms. Simulated cost comparisons suggest meaningful efficiency improvements compared to traditional labeling pipelines. From a business standpoint, the project demonstrates scalable annotation design, human in the loop integration, and structured supervision strategies aimed at balancing cost, speed, and precision in ML development.

Full Technical Stack
AnthropicPythonActive LearningFastAPIReactPostgreSQL
Back to Overview