Works

NLP / LLM

MCP Walkthrough

MCP Walkthrough

An MCP-powered research assistant — from basic tool calling to a multi-server chatbot with resources, prompts, and Claude Desktop integration.

Tokenization Embeddings Attention GPT-2 · LLaMA 3 · Qwen 3 Training + SFT Generation

LegoLLM

Ongoing

Complete LLM framework from first principles — tokenization, attention, training, generation, alignment as modular Lego pieces.

LangGraph Research Assistant

LangGraph Research Assistant

Multi-agent research system with parallel interviews, map-reduce synthesis, and human-in-the-loop approval.

Task mAIstro

Task mAIstro

AI task manager with persistent memory (profile, todos, instructions) deployed via Docker Compose with Postgres and Redis.

RAG Pipeline Architecture

RAG from Scratch

Minimal RAG pipeline built from scratch with NumPy — no framework abstractions.

PDF LangChain Split + Embed ChromaDB Results Query → Semantic Match → Passages

Semantic Search with LangChain

Natural language querying over PDFs using LangChain, ChromaDB, and HuggingFace embeddings.

Agent vs Chain Patterns User Query Agent (decides) retrieve? → generate Chain (always) retrieve → generate

LangChain RAG Patterns: Agent vs. Chain

Comparing LangChain Agent and Chain RAG patterns with multi-provider support.

LangGraph + Human-in-the-Loop Auth LangGraph Agent Human Approval Send authenticate · check_inbox · send_email

LangChain Email Agent

AI email agent with auth-based tool access and human-in-the-loop approval via LangGraph.

                    graph LR
                        A[SmolLM] --> B[LoRA Fine-Tuning]
                        B --> C[AI-as-a-Judge Eval]
                        

SmolLM Fine-tuning

Fine-tuning HuggingFace's SmolLM model on a toy instruction dataset with LoRA and AI-as-a-judge evaluation.

Gemini Long Context

Gemini Long Context

Endangered language translation and codebase security analysis using Gemini 1.5 Pro's long-context window.

LoRA Insights

LoRA Insights: PEFT Recipes

Reproducing and extending LoRA/QLoRA experiments across Llama 3.2 and Qwen 2.5 on H100 GPU.

PEFT-FLANT5

PEFT-FLANT5

Parameter Efficient Fine-Tuning of FLAN-T5 model for Dialogue Summarization.

                    graph LR
                        A[FLAN-T5] --> B[PEFT Fine-Tuning]
                        B --> C[PPO Training]
                        C --> D[Detoxified Model]
                        

PEFT-FLANT5-Detoxification

Fine-Tune FLAN-T5 with Reinforcement Learning (PPO) and PEFT to Generate Less-Toxic Summaries

Captcha-OCR

Captcha-OCR

OCR in captchas with Keras


AI Engineering / MLOps


Computer Vision


Open Source


Other