I have 3 iOS apps on the App Store. I built all of them with AI. I'm currently working on my fourth — a RAG engine that runs completely on-device. I build because I genuinely can't turn this thing off.
I wanted an iOS client for OpenAI's Assistants API. OpenAI never made one. So I figured it out and built OpenAssistant.
Then I wanted to do RAG with Pinecone on my phone. So I built OpenCone.
Then Assistants got deprecated and I needed the new Responses API with all the tools. So I built OpenResponses. It passed App Review on first submission.
Now I'm working on OpenIntelligence — hybrid search, on-device embeddings, Apple Intelligence routing, no cloud calls unless you explicitly consent. It's the most complex thing I've built and I'm genuinely fascinated by it. It runs offline.
I work full-time as a medical device specialist at the VA in Palo Alto. I build at night and on weekends because this is what my brain wants to do. I've figured things out this far, and I trust I'll keep figuring them out as I go.
I see a problem, I figure out how it should work, and I don't stop until it's on the App Store. The code comes from models—the vision is mine.
The successor to my Assistants-era client, rebuilt as an iOS OpenAI Responses API playground with AppContainer dependency injection, SwiftUI, and MVVM handling Responses streaming events.
computer-use-preview, with cancellable runs, reasoning traces, and per-conversation parameters.OpenAIService streams token counts, tool execution cards, and status chips across iPhone and iPad.An on-device RAG engine (RAGMLCore) built for iOS with SemanticChunker, NLEmbedding, and hybrid retrieval fused through RRF + MMR, all wrapped in a consent-aware SwiftUI interface.
NLEmbedding.ContainerService, RAGService, HybridSearchService, and PersistentVectorDatabase coordinate ingestion, retrieval, and storage per isolated KnowledgeContainer.A shipped iOS knowledge base that ingests local documents, builds Pinecone serverless indexes, and streams OpenAI Responses answers with inline citations.
The original Assistants API (v2) client that managed assistants, vector stores, and tool stacks before the Responses API shipped; now maintained as a legacy reference.
Leveraging LLMs to build tools. Since my first commit in late 2023, I've used Foundation Models to generate, refine, and ship complex native iOS applications. Driven by a passion for exploration, I orchestrate systems involving RAG, Agents, and on-device intelligence to close the gaps I see in existing software.
Sole technical specialist supporting Stanford surgical teams with complete autonomy. Bridging the gap between complex medical technology and clinical workflows in high-pressure surgical environments.
I'm always interested in new opportunities and collaborations, especially in iOS development and AI integration. Feel free to reach out if you'd like to discuss a project, explore potential partnerships, or just say hello!