Back to blog
EngineeringMar 21, 2026· min read

Teaching Machines to Know What They Don't Know

We just shipped the first phase of self-improvement for Strug Works: the ability to detect capability gaps through structural honesty.

Most AI systems fail silently. They hallucinate, confabulate, or simply produce confident nonsense when faced with tasks beyond their capabilities. They don't know what they don't know.

Today we're changing that for Strug Works.

What Shipped

Phase 1 of our self-improvement initiative introduces structural honesty and gap detection. The system can now:

• Recognize when a task requires capabilities it doesn't possess
• Identify missing tools, context, or knowledge
• Surface these gaps explicitly rather than attempting to work around them
• Log capability boundaries for analysis and improvement

This isn't self-healing yet. Agents won't autonomously learn new skills or request new tools. But they will tell you honestly when they can't complete a task and why.

Why It Matters

Structural honesty is the foundation for everything else. You can't build self-improving systems without accurate self-assessment. An agent that doesn't know its limits will waste time on impossible tasks or, worse, deliver broken work with high confidence.

For technical leaders using Strug Works, this means fewer silent failures and more actionable feedback. When an agent reports a gap, you know exactly what's missing—whether it's a missing API integration, insufficient context, or a task that requires human judgment.

It also means we can start collecting real data on capability boundaries. Every identified gap is a data point for improving the platform.

How It Works

Agents now perform pre-task validation against their available tools and context. When they encounter a requirement they can't satisfy, they report it through a structured gap detection protocol rather than attempting workarounds.

The implementation is intentionally conservative. We'd rather have agents surface false gaps (things they could handle but flag as uncertain) than miss real ones. Precision will improve as the system learns its actual boundaries.

What's Next

Phase 2 will add remediation capabilities. When an agent identifies a gap, it will be able to:

• Request new tools or integrations
• Ask clarifying questions to fill knowledge gaps
• Propose alternative approaches that work within current capabilities
• Learn from successful gap resolutions

Phase 3 will close the loop: autonomous skill acquisition based on frequently encountered gaps. If agents consistently need a particular capability, the system should learn to provide it.

For now, we're focused on detection. Teaching machines to know what they don't know is hard enough.