Hacker Newsnew | past | comments | ask | show | jobs | submit | chrisford's commentslogin

The vision model has consistently been degraded since 3.5, specifically around OCR, so I hope it has improved with Claude Sonnet 4.5!


Why?


ChatGPT or Claude can't review thousands of medical records at once.


It finds details that are hard for humans to see. It won't get fatigued when looking through thousands of records and stays consistent.


Similar markets, yes.


Why? We are just another AI agent trying to help humans automate tedious work.


Specifically, because you will require vast human health records to train your model, and that model will interact with my health records, and I trust you just about as far as I can throw you as a steward of my or the public’s data. You will intentionally or accidentally expose me to risk, with no meaningful punishment.

Now as a person, of course I’m sure you are kind and responsible and we’d have a lovely lunch (and I mean that). It sounds like a fascinating problem to solve. As a group though, acting within a regulatory regime that doesn’t value privacy at all - excepting that one law from 1996 with more holes than my socks - you just can’t be trusted.

Would you claim personal responsibility for any downside risks your product introduces, in a “like for like” manner with respect to the actual damage caused? Like if a doctor relying on your product caused a death?


No, everything is human-monitored. There is a startup working on AI arbitration though - https://judgeai.space/


It's a productivity AI agent meant for workers who were already dealing with medical records and creating medical chronologies manually, primarily in the legal space.


This is correct. Also, when a law firm client goes under contract with a firm, they often sign HIPAA-compliant medical-release documentation.

Some software providers in the AI medical space are conservative here and have customers sign a BAA directly on signup.


Yes, we are not using the dreaded RAG to accomplish this task, although we do have it available if a customer wants to chat with their case documents.


I get why you’d want to distinguish yourself from competition that relies heavily on RAG, but “chatting” is putting it mildly.

I’d use RAG with prompts like “what was billed but did not produce a record”. Or rehydrating the context of, for example, the hospital’s own model for predicted drug interactions. I could see it being lucrative if that model produced those results without traceability.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: