Using AI to analyze health data has such a huge potential upside, but it has to be done locally.
I use [insert LLM provider here] all the time to ask generic, health-related questions but I’m careful about what I disclose and how I disclose it to the models. I would never connect data from my primary care’s EHR system directly to one of these providers.
That said, it’ll be interesting to see how the general population responds to this and whether they embrace it or have some skepticism.
I’m not confident we’ll have powerful/efficient enough on-device models to build this before people start adopting the SaaS-based AI health solutions.
ChatGPT’s target market is very clearly the average consumer who may not necessarily care what they do with their data.
The last bit
> supervised by a human who occasionally knew what he was doing.
seems in jest but I could be wrong. If omitted or flagged as actual sarcasm I would feel a lot better about the project overall. As long as you’re auditing the LLM’s outputs and doing a decent code review I think it’s reasonable to trust this tool during incidents.
I’ll admit I did go straight to the end of the readme to look for this exact statement. I appreciate they chose to disclose.