Hacker News new | past | comments | ask | show | jobs | submit login
Show HN: API Testing and Security with AI (qodex.ai)
8 points by siddhant_mohan 49 days ago | hide | past | favorite | 3 comments



It claims to plug into your CI pipeline, detect what changed, and generate relevant test cases using LLMs.

As someone who’s struggled with stale or missing tests—especially in fast-moving codebases—I find this idea quite compelling. But I’m also curious about how it handles:

Contextual understanding across large codebases (e.g., multiple modules touched in a PR) Avoiding flaky or non-deterministic tests Matching team-specific coding styles or conventions


Generating tests is good, but how to handle the updating tests? Also how will you handle the flakiness and side effects of AI models?


We handles flakiness with retries, smart waits, and isolation, while side effects are avoided using clean setups, teardowns, and state-safe mocks. Each tests scenarios are independent of each other and can be configured in a way to have prerequisite to setup the system and the post callback to cleanup the system

About updating test scenarios, we map it with your github commits and when a new commits come, we use the diff to figure out if tests failing are because of a bug or because of a new feature.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: