
"He connected the services to his Apple Health account, granting them visibility into long-term fitness and activity records, and later allowed access to select medical information to see how each system handled deeper context. Fowler asked straightforward, consumer-style questions about overall cardiovascular health, the kind a typical user might pose when trying to make sense of years of tracked data."
"When the reporter asked both systems to grade his cardiovascular health, ChatGPT returned failing marks, at one point assigning an F, while Claude issued a more moderate C-range score. Despite the difference, both chatbots framed their responses as meaningful summaries of long-term heart health rather than as limited or uncertain estimates. Those grades conflicted with a physician's assessment of the same data."
A user enrolled in limited-access ChatGPT Health and Anthropic Claude programs and connected both to an Apple Health account containing years of Apple Watch fitness and activity records. The user later granted access to select medical information and asked consumer-style questions about overall cardiovascular health. ChatGPT produced failing grades (including an F) while Claude returned a C-range score. Both systems presented their outputs as meaningful summaries and relied heavily on Apple Watch fitness estimates. A reviewing physician, using the full medical record, found no cause for concern.
Read at TechRepublic
Unable to calculate read time
Collection
[
|
...
]