DrugQ
Overview
DrugQ is an Android research prototype for a multimodal OTC drug safety assistant. It supports voice and image inputs to help users make safer medication decisions, then returns an easy-to-understand safety verdict with an explanation.
Scope
- Implemented as a working Android prototype (not only a concept).
- Initial medication set: Tylenol, Aspirin, Benadryl, and Rolaids.
- Primary symptom contexts: common everyday cases such as cold/flu-like symptoms and allergies.
Problem
- OTC medication guidance is often unclear for non-experts.
- Users may not know how to interpret warnings, contraindications, or drug interactions.
- When users feel unwell, they need a fast, low-effort way to get a safety-oriented answer.
System Design
Inputs
- Voice: users speak the drug name and describe symptoms. Voice is transcribed via OpenAI Whisper.
- Image: users capture photos of the medication (package/label or pill). The image is processed using Gemini + prompt-based interpretation.
Reasoning and Decision
- The system uses an LLM-only approach (no separate rule engine) with prompt design to produce a final safety verdict.
Outputs
- The assistant returns one of three outcomes:
- Do not use
- Consult a doctor
- Safe for use
- Each outcome includes a short, user-facing explanation based on the study’s questionnaire-driven criteria.
What I Built / Contributed
- Built the full Android app end-to-end using Kotlin + Jetpack Compose.
- Implemented the multimodal pipeline (voice + image) and integrated the model calls.
- Designed and iterated prompts to make outputs consistent with the three-category verdict system.
- Conducted AI integration testing and refined the interaction flow based on lab feedback.
Tech Stack / Tools
- Android (Kotlin, Jetpack Compose)
- Firebase (including Firebase AI Logic for on-device API calls)
- Gemini (prompt-based reasoning for image and decision)
- OpenAI Whisper (speech-to-text)
- Figma (flow prototyping)
Collaboration
- Conducted as a lab study project with Prof. Eun-Kyung Choe’s team (University of Maryland).
- I owned the full engineering scope: mobile implementation, AI integration, and prompt design.
Outcome
A working multimodal prototype that produces a clear three-level safety verdict (“Do not use / Consult a doctor / Safe for use”) with explanations. The prototype was developed and iterated within a lab-study setting.