PortfolioΒ 

Watch my comprehensive portfolio showcase videos featuring my latest works in AI & Machine Learning, Full-Stack Development, and Modern Web Technologies

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

Agentic AI Project

MAG7 SEC Filings Analyzer β€” Agentic Financial Intelligence Platform

OPENAI / ANTHROPIC / OLLAMADETERMINISTIC ROUTERFAST RAG (SINGLE-CALL)SEC (10-K / 10-Q)CITATIONSPINECONE (SERVERLESS)SENTENCE-TRANSFORMERSHYBRID RETRIEVALRERANKINGQUERY REWRITESEMANTIC CACHING (MD5)MULTI-PROVIDER LLMFASTAPIREACT 18

This is not just RAG. It is a controlled, production-shaped system for answering questions over real SEC 10-K / 10-Q filings with VERIFIABLE CITATIONS.

The workflow is intentionally ULTRA OPTIMIZED for latency + cost: a DETERMINISTIC ROUTER classifies intent (single-company, comparison, ingestion) with NO LLM CALL, then a FAST RAG AGENT compresses retrieval + analysis + reporting into ONE LLM call (about 3Γ— fewer calls than traditional chains).

Example questions:

  • "Compare AAPL vs MSFT: biggest risks and how they differ."
  • "What were the biggest YoY changes in operating margin, and why?"
  • "What changed from last quarter and why?"
  • "Extract and summarize all risk factors from the latest 10-K."
  • "Identify key growth drivers mentioned in management commentary."
  • "Analyze revenue trends and segment performance across quarters."

Retrieval runs on PINECONE (SERVERLESS) with sentence-transformer embeddings, and advanced options of HYBRID RETRIEVAL, SECTION BOOSTING, RERANKING + QUERY REWRITING. An MD5-KEYED SEMANTIC CACHE returns repeated queries in ~20ms.

Ships with a MULTI-PROVIDER LLM abstraction (OpenAI / Anthropic / Ollama) with pooled + cached instances, plus performance engineering via async FastAPI, request deduplication, and concurrent comparison execution.

πŸ“Š Real SEC Filings

Ingests and indexes actual 10-K / 10-Q filings from all MAG7 companies β€” answers are grounded in real financial documents with verifiable citations.

⚑ 3Γ— Fewer LLM Calls

Deterministic router classifies intent with zero LLM overhead, then a Fast RAG Agent compresses retrieval + analysis + reporting into a single call.

πŸ” Hybrid Retrieval

Semantic + keyword search with section boosting and Cross-Encoder reranking β€” surfaces the most relevant filing passages even for complex multi-part questions.

πŸš€ ~20ms Cache Hits

MD5-keyed semantic cache returns identical or near-identical queries almost instantly β€” dramatically reducing cost and latency for repeated analysis.

πŸ€– Multi-Provider LLM

Pooled + cached LLM instances across OpenAI, Anthropic, and Ollama β€” swap providers freely with concurrent comparison execution for benchmarking.

πŸ“ˆ Company Comparison

Ask cross-company questions like "Compare AAPL vs MSFT risks" β€” the agent routes to a parallel comparison pipeline and merges results into a unified analysis.

πŸ—οΈ Infrastructure

ContainerizationDocker + Docker Compose
App HostingAWS EC2
Edge / HTTPSAWS CloudFront
Reverse ProxyNginx
IaCPulumi (TypeScript)

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

Agentic AI - RAG App Project

Policy Guardian β€” Enterprise Policy RAG + Citations + Streaming

RAGCITATIONSSTREAMINGMULTIMODALQUERY EXPANSIONHYBRID SEARCHRERANKINGCROSS-ENCODERAUTO REWRITELANGGRAPHLANGCHAINPINECONEREDISFASTAPIMULTI-PROVIDER LLMSSEPOSTGRESQL AUDIT LOGSDOCKER COMPOSE

πŸ“‚ Upload any contract, policy document, or regulatory filing and the agent will answer your questions with answers grounded directly in the content of your documents β€” no hallucinations, no guesswork. Every response is backed by exact citations with page numbers so you can verify every claim instantly.

πŸ–ΌοΈ You can also upload image files β€” photos, scans, or screenshots β€” and run them against your loaded policy documents. Ask real-world comparison questions like: "Is this damaged suitcase eligible for a refund under our baggage reimbursement policy?" The agent visually analyzes the image and cross-references it against the relevant policy clauses to give you a grounded, citation-backed answer.

🧠 Powered by a custom fine-tuned model β€” purpose-built and specialized on contract, policy, and regulatory document language. Unlike generic LLMs, this model has been trained specifically for compliance-style Q&A, resulting in significantly higher accuracy, more precise clause interpretation, and faster, more consistent responses compared to out-of-the-box models on the same domain.

A production-style policy compliance assistant that answers questions from internal policy documents with REAL CITATIONS, REAL-TIME STREAMING, and optional MULTIMODAL (text + images) support using CLIP library.

Example questions the app can answer:

  • "Is this damaged baggage eligible for refund under our travel reimbursement policy?"
  • "If an employee shares customer data externally by mistake, what steps must be taken within the first 24 hours?"
  • "Compare our Data Retention Policy vs Vendor Security Policy β€” where do responsibilities overlap?"

The app returns:

  • βœ… Confidence score
  • πŸ“„ Exact citations with page numbers
  • πŸ”— Reference documents
  • 🎯 Full traceability and verification

The RAG pipeline is orchestrated with LANGGRAPH + LANGCHAIN and follows a clear flow (query embedding β†’ vector search β†’ top-k retrieval β†’ context assembly β†’ LLM generation β†’ citation extraction). Retrieval runs on PINECONE (SERVERLESS) with semantic search + metadata filtering, enhanced with QUERY EXPANSION, HYBRID SEARCH (semantic + keyword), and relevance boosting via RERANKING using a CROSS-ENCODER(plus optional AUTO REWRITE to refine user queries). A REDIS caching layer reduces repeated retrieval/LLM calls for faster, more consistent answers. Supports a unified MULTI-PROVIDER interface (Ollama, OpenAI, Anthropic) and streams token-by-token with SSE. Every answer includes citations + is recorded to POSTGRESQL AUDIT LOGS for traceability; shipped with a production-friendly DOCKER COMPOSE setup.

πŸ“„ Cited Answers

Every response includes exact citations with page numbers and reference documents β€” no hallucinations, full traceability back to the source policy.

⚑ Real-Time Streaming

Token-by-token SSE streaming delivers answers instantly as the LLM generates them β€” no waiting for the full response to complete.

πŸ” Hybrid Retrieval

Combines semantic vector search (Pinecone) with keyword matching, reranked by a Cross-Encoder for maximum relevance β€” plus optional auto query rewriting.

πŸ€– Multi-Provider LLM

Unified abstraction supports OpenAI, Anthropic, and Ollama (local) β€” swap providers without changing any application logic.

πŸ—„οΈ PostgreSQL Audit Logs

Every query and answer is recorded to a persistent audit log β€” enabling compliance reporting, usage tracking, and full answer verification.

πŸ–ΌοΈ Multimodal Support

Optional image understanding via CLIP library β€” ask questions about policy diagrams, charts, and visual content embedded in documents.

πŸ—οΈ Infrastructure

ContainerizationDocker + Docker Compose
App HostingAWS EC2
Edge / HTTPSAWS CloudFront
Reverse ProxyNginx
IaCPulumi (TypeScript)

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

Agentic AI Project

Agentic AI Travel Planner

MULTI-AGENTLANGGRAPHSUPERVISORFASTAPIAWS LAMBDAAMADEUS APIPOSTGRESQL (NEON)PRISMATANSTACK QUERYLANGSMITH

An agentic AI travel-planning platform that generates fully personalized itineraries using a MULTI-AGENT system orchestrated by LANGGRAPH. A SUPERVISOR agent analyzes user intent and delegates work across SPECIALIZED AGENTS (RESEARCH, LOGISTICS, COMPLIANCE, EXPERIENCE)to produce structured, consistent plans.

Users can:

  • Choose any city or location
  • Set trip duration (e.g., 3 days, 7 days, etc.)
  • Define preferences (budget, food, culture, adventure, etc.)
  • Instantly receive a structured daily plan

The AI generates:

  • πŸ“ Exact place names
  • 🏠 Real addresses
  • πŸ—Ί Logical day-by-day schedules
  • 🍽 Recommended restaurants
  • 🎯 Attractions and experiences
  • ⏱ Organized timeline throughout the day

Plans are enriched using real travel data via the Amadeus API to ensure locations are accurate and realistic.

πŸ—ΊοΈ Google Maps Route

Interactive day-by-day route map with numbered, color-coded pins for each stop β€” attractions, restaurants, and landmarks plotted in logical travel order.

πŸ’° Estimated Costs

Full trip cost breakdown β€” accommodation, food, activities & transport β€” with a total estimate so travelers can plan their budget at a glance.

πŸ’‘ Local Tips

AI-curated insider tips for each city β€” transport hacks, free entry days, local markets, and hidden gems recommended by the agent.

🏨 Recommended Hotels

Curated hotel picks with star ratings, price ranges, addresses, and descriptions β€” spanning budget to luxury, tailored to the destination.

🍽️ Dining / Restaurant Recommendations

Breakfast, lunch & dinner recommendations with cuisine type, price tier, and exact addresses (and shown on the google Map) β€” so every meal is part of the experience.

The orchestration runs in a separate FASTAPI service deployed on AWS LAMBDA for stateless, pay-per-use scaling, enriched with real travel data via AMADEUS API. Trips persist to POSTGRESQL (NEON) with PRISMA, while the Next.js UI supports planning, saving, and reviewing across devices. Reliability comes from retries + structured flows and LANGSMITH debugging.

Tech: Next.js β€’ React 18 β€’ FastAPI β€’ AWS Lambda β€’ LangGraph β€’ LangChain β€’ OpenAI β€’ Postgre SQL (Neon) β€’ Prisma β€’ Tailwind CSS β€’ Clerk β€’ TanStack Query β€’ Amadeus API

πŸ—οΈ Infrastructure

ContainerizationDocker + Docker Compose
App HostingAWS EC2
Edge / HTTPSAWS CloudFront
Reverse ProxyNginx
IaCPulumi (TypeScript)

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

Agentic AI E-Commerce (MCP + RAG)

Tweeky-Queeky Shop β€” Agentic AI E-Commerce (Agent Gateway + MCP + RAG)

AGENT GATEWAYPINECONE + EMBEDDINGSONE /CHAT SURFACEINTENT ROUTINGMCP TOOL SERVERTOOL USE (PRODUCTS/ORDERS)RAG PIPELINE (TF-IDF)FASTAPIREACT 18REDUX TOOLKITRTK QUERYDOCKERBEANIE + MONGODBJWTPAYPALSTRIPEDOCKERIZEDHYBRID SEARCH

Enterprise Scale project intentionally built to showcase AGENTIC AI patterns inside a real e-commerce app. The React frontend talks to an AGENT GATEWAY through a single chat surface. The agent performs INTENT ROUTING and delegates to either an internal MCP TOOL SERVER (explicit tools for product search/lookup + order tracking) or a RAG SERVICE (local docs retrieval for policy/support Q&A).

Instead of traditional filters, users can simply ask:

  • "Find me products between $200–$300 with the best ratings."
  • "What microphones are best for podcasting?"
  • "Find chairs shorter than 40 inches and under 5kg."
  • "Track my latest order."
  • "What's your return policy for electronics?"

This is a production-style full-stack e-commerce application built using React, FASTAPI, MONGODB, and DOCKER. It provides a complete end-to-end shopping experienceβ€”from user authentication and product browsing to secure checkout and order trackingβ€”all backed by scalable async REST APIs and ROLE-BASED admin management.

The system includes a fully functional admin dashboard for managing users, products, orders, and inventory, along with secure JWT-based authentication, image upload support, and PAYPAL + STRIPE payment integration. πŸ›‘οΈ The frontend communicates with a containerized backend API and MongoDB database, forming a deployable, cloud-ready full-stack solution.

The system understands intent and automatically chooses the correct backend action. πŸ€– The key production-shaped boundary is that the UI never calls TOOLS or RETRIEVAL directlyβ€”everything stays behind the GATEWAYβ€”making the tool-use layer AUDITABLE and TESTABLE, while enabling optional HYBRID SEARCH (BM25 + embeddings) via PINECONE/OPENAI.

πŸ”§ What You Can Do with This App

  • You can create an account, securely log in, and manage your profile.
  • You can browse products, add them to your cart, and complete checkout.
  • You can make secure payments using PayPal and Stripe.
  • You can view your past orders and track delivery status.
  • You can use the admin dashboard to create products, upload images, manage inventory, and process customer orders.
Fine-Tuning Project

Fine-Tuning with Ollama + QLoRA β€” Policy Compliance LLM

QLORAOLLAMALLAMA 3.1 8B4-BIT NF4PEFTHUGGINGFACE TRANSFORMERSBITSANDBYTESGGUFEVALUATION HARNESSPYTHON

A reproducible, end-to-end fine-tuning pipeline for building a domain-specialized POLICY COMPLIANCE model using QLORA (4-bit quantization), then deploying it for local inference with OLLAMA. Built for repeatable experiments and real deployment, from dataset β†’ fine-tune β†’ packaging.

Implements parameter-efficient fine-tuning of LLAMA 3.1 8B with 4-BIT NF4 quantizationβ€”enabling training on a SINGLE GPU. Policy answer accuracy improved by about POLICY ANSWER ACCURACY IMPROVED BY ABOUT ~70% versus the base model, with strong convergence (training loss reduced by ~79% over 3 EPOCHS). Includes an automated EVALUATION HARNESS, a clean END-TO-END PIPELINE (data generation β†’ fine-tuning β†’ adapter merge β†’ evaluation β†’ packaging), and config-driven experiments with clean artifact separation. Ships a merged model packaged for OLLAMA (adapter merge β†’ GGUF) to support fast, OFFLINE INFERENCE.

Tech: Python β€’ HuggingFace Transformers β€’ PEFT (LoRA / QLoRA) β€’ BitsAndBytes β€’ Accelerate β€’ Ollama β€’ GGUF

Fine-Tuning Project

GPT-4 Fine-Tuning β€” Enterprise Policy Compliance AI

FINE-TUNINGGPT-4O-MINIOPENAI APIPYTHONJSONLBENCHMARKINGTOKEN OPTIMIZATIONLATENCY ANALYSISMLOPSAUTOMATION

A production-ready FINE-TUNING and evaluation pipeline for GPT-4O-MINI, designed to improve enterprise policy compliance accuracy while reducing TOKEN USAGE and LATENCY. The domain-specialized model trained on 164 curated JSONL examples boosted accuracy from 35% β†’ 77.5% (~2.2Γ— better) on compliance-focused evaluation sets. Includes an AUTOMATED EVALUATION SUITE with CATEGORY-LEVEL SCORING, REGRESSION CHECKS, and performance reportsβ€”token usage dropped by ~20%, and response latency improved from 1.61s β†’ 1.06s (~44% faster). The pipeline follows an MLOPS-STYLE WORKFLOW with training data versioning, job automation, and model validation in OPENAI PLAYGROUND.

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

Serverless Project

Serverless Order Management System β€” Event-Driven Cloud Platform

SERVERLESSAWS LAMBDAAPI GATEWAYSQSSNSDYNAMODBEVENTBRIDGESTEP FUNCTIONSAWS CDKTYPESCRIPTREACTCHART.JS

A fully SERVERLESS, EVENT-DRIVEN Order Management System on AWS, built for high scalability, operational visibility, and real-time analytics. Order workflows run through AWS LAMBDA behind API GATEWAY, with decoupled processing via SQS, SNS, and DYNAMODB.

Complex lifecycle flows are coordinated with STEP FUNCTIONS (retries, failure paths, DLQs) and optional EVENTBRIDGE pipelines. The whole environment is provisioned with AWS CDK (TypeScript) + CloudFormation, and shipped with a modern REACT + TYPESCRIPT dashboard featuring analytics charts (Chart.js) and CSV import/export workflows. Security and observability lean on least-privilege IAM + CloudWatch monitoring.

Tech: AWS Lambda β€’ API Gateway β€’ SQS β€’ SNS β€’ DynamoDB β€’ S3 β€’ Step Functions β€’ EventBridge β€’ AWS CDK β€’ React 19 β€’ TypeScript β€’ Chart.js

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

Full-Stack Project

Cocktail Maker App - Serverless Lambda + React Query + PostgreSQL

Mixmaster is a full-stack cocktail discovery app built with React βš›οΈ, React Query ⚑, Node.js 🌐, and PostgreSQL πŸ—„οΈ. It blends real-time data from TheCocktailDB 🍸 with custom user-created recipes stored locally with uploading your image to S3 bucketπŸͺ£, enhanced by intelligent caching for high-speed performance. A serverless AWS Lambda microservice ☁️ handles newsletter automation with Amazon SES βœ‰οΈ, delivering a smooth, modern UI and fast, reliable CRUD operations.

πŸ’‘ Tip: Increase video quality settings for the best viewing experience

E-Commerce Website

Tweeky Queeky Shop (MERN)

A fully Dockerized MERN e-commerce platform built with React βš›οΈ + Redux Toolkit 🧩, Node/Express πŸš€, and MongoDB πŸ—„οΈ. It delivers a real-time product experience with secure JWT authentication πŸ”, PayPal/credit-card checkout πŸ’³, image uploads πŸ“Έ, and full admin management workflows. Designed with a scalable, production-ready architecture ideal for real-world store operations.