Architected a dual-transformer system pairing DeBERTa-v3-Large (434M) and RoBERTa-Large (355M) with agreement-based routing; LLM arbitration triggered only when models disagree (~1% of queries), reframing the LLM's task from ternary to binary classification. Avoids the failure mode of confidence-based routing on overconfident model errors. Key finding: architecture matters more than scale, and inter-model disagreement is a better uncertainty signal than softmax confidence. Presented at Johns Hopkins Carey Business School national championship, February 2026.
Eight years shipping production systems — healthcare, then Meta.
Engineer in Austin, TX. Built a COVID warehouse spanning 46 hospitals at Prime, scaled NLP automation touching tens of millions of cases at Meta, and won Data4Good 2025 with a dual-transformer hallucination detector. Day job and research keep informing each other.
Selected work
Independent research investigating whether persona framing like "You are an expert in X" — a norm borrowed from human collaboration contexts — imposes an unnecessary scarcity constraint on LLMs. Core hypothesis: telling an LLM to be a specialist doesn't activate specialized knowledge; it artificially suppresses everything else, potentially narrowing the conceptual search space on generative ideation. Refined task-dependent hypothesis: expertise framing is neutral-to-noise on retrieval-heavy factual tasks, beneficial for structured reasoning, potentially harmful for creative/generative ideation. Measurement pipeline uses Shannon entropy over a 15-domain taxonomy.
Structured decision-making tool: AI personas (optimist, pessimist, analyst, creative, synthesizer) offer distinct perspectives in parallel while the user provides emotional context as the Red Hat. The Blue Hat (synthesizer) integrates all inputs including the user's own. Underlying data model supports custom personas, custom panels, weighted ensembles, and alternate deliberation frameworks beyond Six Hats. Provider-abstracted AI layer supporting both Anthropic Claude and OpenAI GPT-4o.
Open-source Python library for confusion-matrix analysis and ML meta-research auditing. Features include metric reconstruction, probabilistic inference, statistical testing, and cost-sensitive analysis. Designed to catch errors and implausible results in published ML evaluations. Reflects an applied interest in ML reliability and the integrity of published evaluation metrics.
Built and maintained a COVID data warehouse aggregating data from all 46 Prime Healthcare hospitals to support corporate decision-making and government reporting. Developed automated report interfaces and data feeds; surfaced proactive savings opportunities and value analyses for contracts under review. Most of the original architectural work I did pre-pandemic was compressed into a few months when the warehouse became operational under time pressure.
Experience
NLP automation and end-to-end support journeys at platform scale.
- ▸Own end-to-end user support journeys for global developer + business surfaces; design intake funnels between self-service and assisted channels.
- ▸Drive reliability and DX metrics (CSAT, re-open rate, TTR) via debugging, proactive monitoring, and infra fixes.
- ▸Built NLP-based classification + automation in Hack and JavaScript that triage and auto-resolve cases at Meta scale; LLM-driven intent + issue detection.
- ▸Defined long-term automation strategy and roadmaps; partnered with PMs and analytics to size opportunities and instrument quality metrics.
46-hospital data systems, COVID warehouse, and purchasing automation.
- ▸Built and maintained a COVID data warehouse aggregating data from all 46 Prime hospitals; powered corporate decision-making and government reporting.
- ▸Designed a React portal for physician-preference-item contract administration with Python automation for purchase-approval workflows.
- ▸Automated purchasing across 45+ hospitals in Python — order feeds for recurring items meeting strict pricing criteria; materially reduced manual error.
- ▸Built dashboards surfacing proactive savings opportunities; item crosswalks and contract-analysis tooling for value analyses.
Predictive modeling, stochastic optimization, and four years TAing.
- ▸Predictive modeling and stochastic optimization across multiple stakeholder engagements.
- ▸Optimized data pipelines on cloud infrastructure — significant runtime reductions on production data models.
- ▸TA for analytics courses across four years; consistently top-3 of 16 in the program.
Credentials
- education/
- MS, Business Analytics and Information Management · Purdue University · Krannert · 2018BS, Industrial Management + Economics + Supply Chain Information & Analytics · Purdue University · Krannert · 2017
- publications/
- Detecting AI Hallucinations in Educational Contexts: A Transformer-Based Approach to Protecting Student Learning · Submitted · INFORMS Journal on Data Science · 2026Role of Political Ideology in Friendship Social Networks · Midwest Data Science Conference · 2018Effect of Forecast Accuracy on Inventory Optimization Model · Midwest Data Science Conference · 2018Carrier Choice Optimization with a Tier-Based Rebate Program · Midwest Data Science Conference · 2018
- talks/
- AI Hallucination Detection: A Transformer-Based Approach to Protecting Student Learning · Johns Hopkins Carey · National Championship · 2026Role of Political Ideology in Friendship Social Networks · Midwest DSI · Indianapolis (oral) · 2018Effect of Forecast Accuracy on Inventory Optimization Model · INFORMS Analytics · Baltimore (poster) · 2018Carrier Choice Optimization with a Tier-Based Rebate Program · INFORMS Analytics · Baltimore (poster) · 2018
- awards/
- Data4Good 2025 — National Competition Winner (Non-Academic Track, AI Hallucination Detection) · 2025Best Machine Learning Algorithm — Identification of Clickers from Bookers · Online travel-booking competition · 2017Krannert School of Management Scholarship · $7,500 · Purdue University · 2017
Projects
| Project | Domain | Stack | Year | Status |
|---|---|---|---|---|
LinkedIn Resume Optimizer /linkedin-resume Two-pass LLM pipeline: normalize a PDF resume into structured data, then rewrite it against a job description. | AI · Productivity | FastAPI · React · Vite · Anthropic | 2026 | · private |
Writing
Now
- building/
- Scaling a new generation of LLM support assistants at Meta.Pilot testing Scarcity Fallacy research.
- reading/
- Hitler by Ian KershawLife 3.0 by Max Tegmark