Structure Before Speed: Why Better Decisions Need More Than One Opinion
By Forenta Team
Most bad decisions are not made by people who lacked information. They are made by people who had enough to feel certain, and not enough structure to test whether that certainty was warranted.
This is not a productivity problem. It is a judgment problem. As the cost of building drops, it becomes a more expensive one. When a team can move from idea to working prototype in days, the constraint shifts. The question is no longer whether something can be built quickly. It is whether the right thing is being built, and whether the reasoning behind that choice has been genuinely challenged before work begins.
The structural weakness of single-source analysis
A single analysis, whether from a colleague, a consultant, or an AI assistant, carries a hidden risk that has nothing to do with how capable the source is. It is coherent by construction. One thinker organises information into a logical narrative, and that narrative will always appear more persuasive than the uncertainty it has smoothed over. The gaps are invisible because they were already resolved, or quietly dismissed, by the author.
The cognitive mechanism behind this is well-documented. Tversky and Kahneman showed in a 1974 study in Science that when people estimate, they anchor to an initial value and adjust insufficiently, even when that initial value is arbitrary.¹ When the first analysis in a discussion establishes a frame, subsequent thinking adjusts around that frame rather than generating independent views. This is not a sign of limited intelligence; it is a documented property of how human judgment operates under uncertainty.
At the group level, the effect compounds. Irving Janis's analysis of high-stakes policy failures, from the Bay of Pigs to Pearl Harbour, identified a pattern he called groupthink: cohesive teams under time pressure suppress dissenting views, converge prematurely on shared conclusions, and stop stress-testing their assumptions.² The mechanism is not poor individual judgment. It is the social pressure that emerges when groups need to feel certain about a decision before committing to it.
Concept: Tversky & Kahneman (1974); Janis (1982)
What the research shows about collective judgment
Research on collective intelligence points toward a different model. In a 2010 study in Science, Woolley and colleagues measured group performance across a range of cognitive tasks and found a general factor, a 'c factor,' that predicted group performance better than any individual member's IQ.³ This factor correlated with how evenly participation was distributed and how socially perceptive group members were. It did not correlate strongly with having the most intelligent person in the room.
Francis Galton's 1907 observation at a county fair provides a simpler illustration.⁴ When 800 visitors independently estimated the weight of an ox, the median of all guesses was accurate to within 0.8 percent of the actual value, more accurate than any individual expert. Galton recognised the operative condition: independence. Correlated guesses converge toward shared errors. Independent guesses, carefully aggregated, converge toward accuracy.
Concept: Galton (1907). Vox populi. Nature.
Six decades of structured deliberation
In the early 1950s, the RAND Corporation faced a concrete problem: how to get reliable forecasts from expert groups without distortion by social hierarchy, dominant personalities, or premature consensus. The solution, developed by Dalkey and Helmer and published in Management Science in 1963, became known as the Delphi method.⁵ Experts submit assessments independently and anonymously, receive controlled feedback about the distribution of responses, and revise their views based on the aggregate, not based on who said what.
Roger Cooke's foundational work on structured expert elicitation formalised this further: reliable expert judgment requires calibrating experts against known quantities, weighting contributions by track record, and being explicit about where genuine uncertainty remains rather than forcing false consensus.⁶ The output is not the average of all opinions. It is a calibrated summary of what the available evidence actually supports, including where that evidence runs out.
Gary Klein introduced a complementary technique: the premortem. Instead of asking a team to evaluate whether a plan might fail, you ask them to imagine it has already failed, then identify what went wrong.⁷ Research found that imagining failure as accomplished fact increases the ability to correctly identify failure causes by 30% compared to imagining it as a possibility. The technique works because it legitimises dissent. People who would not openly criticise a plan will freely describe its failure.
Source: Dalkey & Helmer (1963). Management Science.
Why blind rounds improve judgment quality
In a controlled experiment at the 2017 WSDM conference, Tomkins, Zhang, and Heavlin tested what happens when reviewers can see the identity of the authors they are evaluating.¹⁰ Single-blind reviewing produced a significant advantage for papers from prestigious institutions, a bias that disappeared entirely under double-blind conditions. When reviewers knew who wrote a paper, they evaluated the conclusion through the lens of the author's affiliation rather than the quality of the argument itself.
Research on devil's advocacy adds an important nuance. Schwenk's 1990 meta-analysis found that structured dissent, whether devil's advocacy or dialectical inquiry, consistently produces higher-quality decisions than consensus approaches.⁸ But Nemeth and colleagues showed in 2001 that authentic minority dissent is more valuable than any form of role-played devil's advocacy: when a dissenter genuinely holds a different view, it stimulates more divergent thinking than when dissent is assigned as a role.⁹ The goal is not to add a critical voice as a formality. It is to structure the process so that genuine disagreement can surface and survive.
The insight from six decades of expert elicitation research is not that more opinions produce better decisions. It is that independent opinions, collected before group influence takes hold, structured to surface disagreement, and synthesised with explicit attention to uncertainty, produce better calibrated judgment.
From research to product: Council Deliberation
Forenta is developing Council Deliberation as a pre-execution validation layer. A user submits an idea, a product direction, or a strategic question. Six specialised advisors then work in parallel and independently. No advisor sees the others' analysis during the initial round.
After the independent round, a peer review phase asks each advisor to assess the reasoning quality of the others' analyses, not the conclusions, but the argument structure, evidence quality, and unexamined assumptions. A rebuttal pass follows, giving each advisor the opportunity to revise or stand behind their original position. A chairman synthesis then draws the full set of findings into a structured output.
The six advisory frames are designed to ensure no standard analytical angle goes unexamined. The Risk Sentinel examines failure modes and kill conditions. The Ground Truth Architect works from first principles, verifying logical chains and unsubstantiated assumptions. The Opportunity Cartographer maps adjacent possibilities the original framing may have excluded. The Pattern Scout applies reference class reasoning to identify base rates and comparable cases. The Delivery Marshal focuses on build sequence. The Automation Strategist assesses where AI tooling can reduce cost or execution risk.
The output has four components: a findings summary for quick orientation; a detailed advisory report with the full analysis and any dissenting views; a visual HTML infographic designed for sharing and presentation; and a structured JSON output for downstream integration with Forge, builder matching, and project scoping.
What this does not cover
Council Deliberation is designed for decisions that benefit from multiple analytical frames before a commitment is made. It is less useful for decisions that are genuinely reversible at low cost, for execution questions where speed matters more than validation, or for problems requiring specialised domain expertise the system's advisors do not have. The quality of any deliberation is also bounded by the quality of the input: a vague brief produces a less useful output than a specific one.
The research on structured expert elicitation also cautions that more structure does not automatically produce better output. Cooke's work showed that judgment quality depends on calibration and track record within the relevant domain. For novel or rapidly evolving domains, the reference classes that make structured deliberation most valuable may not yet exist.
Judgment is a process, not a moment
The research finding is precise: independent analysis collected before group influence takes hold, structured to surface disagreement, and synthesised with explicit attention to uncertainty produces better-calibrated judgment than consensus-first approaches. This is not a finding about AI or software platforms. It is a finding about how human judgment reliably fails and how that failure is correctable.
The practical implication for founders and builders is straightforward. The point before execution is the cheapest moment to surface a weak assumption. Every week of work built on an unexamined premise adds cost to the eventual correction. Structured deliberation is not overhead. It is what prevents weak reasoning from compounding into expensive problems.
References
- 1.Tversky, A., & Kahneman, D. (1974). Judgment under uncertainty: Heuristics and biases. Science, 185(4157), 1124–1131.
- 2.Janis, I. L. (1982). Groupthink: Psychological Studies of Policy Decisions and Fiascoes (2nd ed.). Houghton Mifflin.
- 3.Woolley, A. W., Chabris, C. F., Pentland, A., Hashmi, N., & Malone, T. W. (2010). Evidence for a collective intelligence factor in the performance of human groups. Science, 330(6004), 686–688.
- 4.Galton, F. (1907). Vox populi. Nature, 75, 450–451.
- 5.Dalkey, N. C., & Helmer, O. (1963). An experimental application of the Delphi method to the use of experts. Management Science, 9(3), 458–467.
- 6.Cooke, R. M. (1991). Experts in Uncertainty: Opinion and Subjective Probability in Science. Oxford University Press.
- 7.Klein, G. (2007). Performing a project premortem. Harvard Business Review, 85(9), 18–19.
- 8.Schwenk, C. R. (1990). Effects of devil's advocacy and dialectical inquiry on decision making: A meta-analysis. Organizational Behavior and Human Decision Processes, 47(1), 161–176.
- 9.Nemeth, C. J., Brown, K., & Rogers, J. (2001). Devil's advocate versus authentic dissent: Stimulating quantity and quality. European Journal of Social Psychology, 31(6), 707–720.
- 10.Tomkins, A., Zhang, M., & Heavlin, W. D. (2017). Reviewer bias in single- versus double-blind peer review. Proceedings of the National Academy of Sciences, 114(48), 12708–12713.