Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

UI detection’s a big focus - we use visual grounding + structured observations (like icons, OCR, app metadata, window state), so the agent can reason more like a user would. It’s surprisingly robust even with layout shifts or new themes


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: