Hacker Newsnew | past | comments | ask | show | jobs | submit | morisil's commentslogin

I've been teaching harness engineering for a while:

https://xemantic.com/ai/workshops/


How do you handle possible prompt injection in emails?

https://github.com/xemantic/markanywhere

Incremental Markdown parser that emits streams of semantic events, plus tools to manipulate them - designed for real-time rendering of streamed LLM output.


Regarding applications/industries - after software development as an industry I see scientific research benefiting the most from agentic AI and autonomous machine reasoning. Another domain which is still not saturated would be something I call "personal guardian angel", which usually extends to whole families. Quite often people use OpenClaw for these use cases. Personally I build my own harnesses. One instance is set up as a CEO of my organization, taking care of operations, and another one, completely separate, operating over private knowledge about my family matters, schedules, medical history, etc. I predict context graph and metacognitive use cases to get rapid adoption this year.

- I've built 2 harnesses, one called Claudine - an older sister of Claude Code which now I use for teaching harness engineering, and Golem XIV, with context/knowledge graph management and self-directed metacognitive research.

- In practice rather Anthropic models if the quality of the metacognitive reasoning process and self-improvement loops are considered.


Thank you


More than a year ago I built my own coding agent called Claudine. I also made agentic anthropic-sdk-kotlin and few other AI libraries for the ecosystem. Maybe this low-level experience allows me to use these tools to deliver in 2 days what would have taken 2 months before.

My advice - embrace TDD. Work with AI on tests, not implementation - your implementation is disposable, to be regenerated, tests fully specify your system through contracts. This is more tricky for UI than for logic. Embracing architectures allowing to test view model in separation might help. I general anything reducing cognitive load during inference time is worth doing.


Suboptimal choice. According to AutoCodeBench, for equivalent problem complexity, LLMs generate correct Kotlin code ~70% of the time versus ~40% for Python, and Go scores lower than Python. Kotlin can be executed as a script while providing super fast compilation phase next to evaluation phase, which is further reducing a chance of mistakes. I don't use tools anymore. I just let my LLMs output Kotlin script directly together with DSLs tailored to the problem space, reducing cognitive load for the machine. It works like a charm as a Claude Code replacement, not only coding autonomously in any language, but directly scripting DB data science, Playwright, etc., while reducing context window bloat.


1. LLMs excell at extracting facts from the context. Storing them as a subject-predicate-object relationships is "natural" for graph databases. Doing it right, so that this knowledge can be utilized more efficiently than any RAG, requires sophisticated context engineering, for example to avoid duplicates and keep consistent vocabulary for relationships, but it is totally achievable and the quality of automatically extracted knowledge can be almost spotless, especially if an LLM can also decide on generating parallel embeddings as a semantic search entry point for graph traversal.

2. Writing cypher queries is a job I would never like to have as a human. But LLMs love it, so that an agent can do an ad hoc data science for every single problem. Especially while being aware which criteria were used for graph construction. It is worth ditching things like MCP in favor of tool graph-like solutions. For this purpose I developed my own DSL which only LLM speaks in internally. The effects are mind-blowing.


I would support every political and social movement progressing us on the spectrum from patriarchy to matriarchy. In particular I would put pressure on the legal system of countries where women are still not allowed to inherit land and property.

The next biggest problems to tackle:

  - the way we are producing proteins
  - the way we are producing energy
Short term problems to address:

  - adoption of cognitive AI in scientific research
I am building very potent autonomous AI agents now, so soon I will be able to unleash them to crunch all these problems, hopefully. :)


> I am building very potent autonomous AI agents now, so soon I will be able to unleash them to crunch all these problems, hopefully. :)

We still need the people that enforces patriarchy over matriarchy to use inferior agents.


Personally I use it for implementing complete new features using test driven development, or fix bugs in existing code base.

Outside of software development I use it for managing all the legal documents of our cooperative (and it's Germany, and I don't speak German well ;) )


It is exactly the opposite. Participants collect all the pieces of knowledge to build an agent like my Claudine at the end of the workshop:

https://github.com/xemantic/claudine/


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: