- I'm beginning to suspect that a key skill in working effectively with coding agents is developing an intuition for when you don't need to closely review every line of code they produce. This feels deeply uncomfortable!Oct 11, 2025 12:31
- Cf. OpenAI “Why LMs Hallucinate.” 70%+ calibration error rates on code generation tasks. Any distributional “intuition” is bound to fail. If we could hone intuition to predict errors, so could the LLMs. If the code matters, you need to look at it carefully.