ML Engineer · LLM Systems · Agentic AI Building production AI systems at scale — LLM-powered agents, RAG pipelines, and backend services that ship.
| Project | What it does | Stack |
|---|---|---|
| llm-knowledge-editing | LLM evaluation platform: RAG ingestion pipelines, pgvector retrieval, MLflow experiment tracking across ROME, LTE-LoRA, and ICE. Run on 4-GPU cluster over 48 hours. | Python, HuggingFace, LLaMA2, LangChain, MLflow |
| dota2-build-generator | Embedding-based recommender trained on 50K+ matches. Deployed as production inference API with TypeScript + Node.js backend, Docker packaging, and CI/CD pipeline. | Python, PyTorch, TypeScript, Node.js, Docker |
| cuda-inlj | GPU-accelerated B+ Tree index nested-loop join. 4.2x speedup over CPU baseline on 10M-row datasets, benchmarked within 5% of theoretical bounds. | CUDA, C++ |
Production LLM agent systems with LangGraph: multi-step tool-use, agentic workflows, and RAG pipelines deployed at scale.
