The Implicit Explorer
The Implicit Explorer
Give an LLM a binary to analyze for vulnerabilities. It does not follow a predefined algorithm. It does not execute a control-flow graph traversal. It reasons — iteratively, over multiple passes — and its reasoning develops structure.
Li, Zhang, and Wang analyze 521 binaries across 99,563 reasoning steps and find four emergent token-level patterns. Early pruning: the model discards unpromising paths quickly, spending attention elsewhere. Path-dependent lock-in: once committed to an analysis direction, the model persists even when evidence weakens. Targeted backtracking: the model reverses selectively, not wholesale. Knowledge-guided prioritization: the model focuses on patterns it recognizes from training, not patterns dictated by the binary’s structure.
These patterns are not programmed. No instruction tells the model to prune early or lock in. The patterns emerge from the interaction between the model’s trained knowledge and the iterative reasoning format. They form a stable, structured system with distinct temporal roles — pruning dominates early passes, lock-in dominates middle passes, backtracking appears late.
The structural finding: LLMs conducting binary analysis have reinvented heuristic search strategies that human reverse engineers use, but without explicit instruction. The convergence is not mysterious — both face the same problem (exponential path space, limited attention budget) and the same constraints select similar solutions. But the path-dependent lock-in is a failure mode that human analysts also exhibit — and unlike human analysts, the LLM cannot recognize it as a bias. The implicit pattern is simultaneously a capability and a vulnerability.