$ whoami
AI Engineer
$ cat bio.txt
4 years of backend engineering in C#/.NET and 2+ years
building production-grade LLM systems — multi-model
routing, RAG pipelines, ReAct agents, and MCP on AWS.
Proven focus on latency optimization and cost efficiency.
$ cat focus.txt
LangChain · LangGraph · CrewAI · MCP · AWS
$ ▋
Multi-Model LLM Routing System
Designed a multi-provider LLM routing layer integrating OpenAI, Groq, Gemini, and Llama with dynamic fallback and routing strategies based on response time and cost constraints. Implements tool calling with web search for real-time knowledge augmentation across text, vision, TTS, and image generation.
Retrieval-Augmented Generation
Built a document Q&A system using LangChain + FAISS with semantic retrieval over 1K–3K document chunks. Designed a ReAct workflow using LangGraph with Tavily fallback, reducing failed queries by ~30–40%; integrated AWS Bedrock AgentCore for persistent memory. Attained ~85% retrieval accuracy at p95 ~1.2s.
Multi-Agent Pipeline
5-agent sequential pipeline (CrewAI) specialized in bug detection, OWASP Top 10 security, performance optimization, and documentation analysis. Processes up to 50 files/run with severity scoring, cutting code review turnaround by ~40%.
MCP-Based AI Assistant
Claude Desktop-integrated study assistant using Model Context Protocol for structured tool interaction with document parsing pipelines (PDF, Markdown). Generates quizzes, flashcards, summaries & explanations — 100+ pages/session.
Conversational Support AI
Stateful conversational AI agent for FAQ-based customer support with persistent memory via AWS Bedrock AgentCore. Semantic search using FAISS + sentence-transformers over ~500–1K FAQ entries; reduced cold-start latency via S3-backed vector store.
Full-Stack Finance App
Full-stack expense tracker with glassmorphism UI, category-based analytics, real-time charts, and 41+ automated test cases. Deployed on AWS Elastic Beanstalk with MySQL RDS.