As a urologist who built and runs his own clinic management software, I'd encourage thinking about this question early: what does the system do when the LLM refuses to answer, returns malformed JSON, or hallucinates a glycemic value? In medical contexts, a 'silent failure' (system continues despite bad data) is much worse than a noisy failure (system stops and asks the user). The 'happy path' for an LLM-powered medical tool is usually well-designed. The failure paths are where the project lives or dies. Curious how you handle that.
[flagged]