Skip to main content
Portfolio
HomeProjectsAboutCertificationsContact

© 2026 Portfolio. All rights reserved.

ContactAdmin

Projects

Case studies and shipped work in LLMs, RAG, and production ML.

RAG pipeline at scale

Reduced retrieval latency by 60% and cost by 40% for a 10M-document knowledge base using hybrid search and custom reranking.

RAGLLM

iOS + on-device AI assistant

Shipped an on-device LLM experience for a productivity app. Optimized for latency and battery; 2M+ monthly active users.

LLMiOS

LLM observability platform

Built an internal platform for tracing, logging, and cost attribution across 50+ model endpoints. Cut debugging time by 70%.

LLMInfra