|
The real performance of a regulatory briefing book is not measured by page count or polish. It is measured by how efficiently a regulatory reader can follow the logic, test it, and reach a defensible position on the sponsor’s ask.
Unfortunately, many briefing books still make the reviewer work too hard. They are organized by how teams built the data—not by how regulators make decisions. That misalignment sounds minor, but it has profound cognitive consequences. The Hidden Cost of Misalignment When a document mirrors internal development logic instead of regulatory reasoning, the reader must reverse-engineer the argument. The regulatory reader sifts through design history (“the largest study ever planned in this patient population”) to locate decision relevance. They assemble fragments of details and data to reconstruct the sponsor’s logic trail. Each mental step adds friction—time, effort, and uncertainty. The concept of friction is not to be underestimated. Legal scholar Brett Frischmann describes friction as resistance or drag built into documents used in legal and regulatory decision processes. Cass Sunstein and Richard Thaler refer to unnecessary complexity in documents as sludge—“bad friction” that drains cognitive energy and delays decisions. In regulated markets, economists have quantified these effects. Studies in The Quarterly Journal of Economics and the American Economic Review show that even modest “choice frictions” in Medicare decisions significantly slow and distort judgments. The parallel for regulatory writing is direct: when reasoning is hard to trace, decision friction rises. I often tell clients that their briefing book delivers information but fails to deliver understanding. Common Patterns of Friction
Tables without interpretive anchors that leave reviewers to construct meaning. Massive appendices meant to prove thoroughness rather than aid reasoning. Each of these choices forces reviewers into reconstruction mode. They also trigger what cognitive-load theorists call the split-attention effect—when information that must be mentally integrated is physically separated across pages, tables, or sections. John Sweller’s research in applied cognitive psychology shows that split attention increases working-memory load, slows comprehension, and makes readers less confident in their conclusions. In briefing books, the effect appears when a data table sits ten pages from its interpretation or when related safety and efficacy evidence live in separate silos. The reader’s mind does the integration work the document should have done. And when reviewers must reconstruct logic, decision friction increases and confidence declines. Reviewers start to question not only the data, but the sponsor’s grasp of its own argument. A document that cannot show its reasoning clearly invites more information requests, longer review cycles, and less trust in sponsor judgment. The Decision Pathway Decision-enabling design exists to solve this exact problem. My vision of information design reorganizes briefing content around the decision pathway—the same mental sequence reviewers follow when testing an argument: Issue → Evidence → Interpretation → Position → Request This sequence is not arbitrary. It mirrors both the FDA Benefit–Risk Framework (issues → evidence → appraisal → conclusion) and the structure used in legal briefs (Issue → Rule → Application → Conclusion). Across disciplines, this logic flow has proven to reduce cognitive friction and increase confidence in reasoning. When that sequence is built into the structure, reviewers stop searching for logic and start evaluating reasoning. That shift—from retrieval to evaluation—is the defining mark of a well-designed briefing book. The Shift: From Information Display to Decision Design Teams often assume that completeness equals clarity. It doesn’t. Regulatory readers don’t reward density—they reward traceable reasoning. A decision-enabling design builds its logic trail around the reviewer’s task. Each section of the book should answer one implicit question: What decision does this content support, and what evidence makes it reasonable? When that sequence is visible, reviewers stay oriented. When it’s hidden, they reconstruct it—usually with lower confidence and more questions. The Hallmarks of Decision-Enabling Design A. Purpose-Anchored Structure Each major section begins with the So what?—why the issue matters and what decision is needed now. This tells reviewers not just what they’re reading, but why it matters. B. Progressive Disclosure Information flows from summary → key evidence → traceable detail. Reviewers can engage at any level of depth without losing sight of the reasoning line. C. Decision Cues at the Point of Use Tables and figures are placed where the decision occurs—not buried in appendices. Each display answers a single regulatory question and closes the loop with interpretation, not adjectives. The Result: Cognitive Efficiency, Not Rhetorical Flourish When design mirrors the decision process, cognitive load drops. Readers no longer have to infer your logic; they can inspect it. Review becomes faster, more confident, and less adversarial. That is the quiet power of decision-enabling design: the approach turns information into defensible reasoning—and does so on the regulatory reader’s terms. The Takeaway A decision-enabling briefing book doesn’t argue harder. It argues smarter. It lets structure do the heavy lifting and keeps reasoning in full view. When regulatory teams master this design discipline, meetings change. Regulatory readers spend less time asking “Where does this fit?” and more time discussing “What does this mean for the development program, the patient and the label?” That’s the point—clarity that enables decision. Join the conversation on LinkedIn: https://www.linkedin.com/pulse/decision-enabling-design-how-briefing-books-should-think-cuppan-3pa7c
0 Comments
Regulatory documents are judged not only by what they say, but by how efficiently the documents help regulatory readers decide. Let’s keep in mind that every document you produce that ends up appearing on the computer screen of a regulatory agency reader is intended to be an advisory document. A document intended to aid the reader in making decisions.
However, my experience suggests that many development teams do not fully appreciate this point. Teams tell me they are in the business of reporting. Teams measure quality by format compliance, data completeness, or adherence to templates — not by decision efficiency. You are not in the reporting business. You are in the business of advising. This judgment gap on “what documents do” explains why many submissions are technically correct but cognitively exhausting to read. Most regulatory documents fail not because they are incomplete, but because they are opaque to reasoning. Regulatory readers are not struggling to find data — they are struggling to follow the thinking behind the data. Every table, paragraph, and conclusion should serve one purpose: to let the reviewer see how evidence informed our thought processes and contribute to decision-making. Regulatory reviewers do not need more information. They need better reasoning access — clear logic, visible connections, and writing that mirrors how decisions are made. Decision efficiency measures how well a document enables a regulatory reader to:
Keep in mind that regulatory readers are auditing reasoning associated with study design, conduct, and data sets. When insight into sponsor reasoning pathways are nonexistent or blocked, the cost is real. This is where the friction comes into play—the reviewer should not have to backtrack, infer, or guess. Reviewers spend time reconstructing logic that writers should have made transparent. Dense text, redundant tables, and poorly signposted arguments force readers to think about the document rather than through it. Each moment of confusion compounds uncertainty. Confidence in the sponsor’s reasoning drops, and the agency’s questions multiply. This is not inefficiency of time — it is inefficiency of thinking. The inefficiency of thinking is the far more dangerous dimension. It erodes trust, clouds the evidence trail, and converts clarity into doubt. In an environment where regulatory review cycles are compressed and AI tools assist human readers, the ability to think clearly through a document — not merely within it — is the new competitive advantage. Decision efficiency evaluates the clarity and logic density of a document — not the volume of its words. At McCulley Cuppan, we define decision efficiency through three observable dimensions:
Why This Metric Matters Now Regulatory authorities worldwide are integrating AI-assisted review tools that rely on structured reasoning and clear metadata. Decision efficiency determines whether those tools (and human reviewers) can interpret content without manual intervention. In this context, poor writing is not just a style flaw — it is an information-access problem. A document with low decision efficiency hides logic behind noise. Moving from Compliance to Cognition Teams that focus only on compliance write to satisfy checklists. Teams that focus on decision efficiency write to support thinking. This shift transforms how review rubrics, templates, and training are used:
This is the next frontier of quality in regulatory writing. Decision efficiency is emerging as the missing metric — the measure that predicts whether a document will accelerate or delay regulatory decision- making. Future articles in this series will show how to design decision-efficient documents, measure them with structured rubrics, and visualize reasoning clarity through AI-enabled analysis dashboards. |
AuthorGregory Cuppan is the Managing Principal of McCulley/Cuppan Inc., a group he co-founded. Mr. Cuppan has spent 30+ years working in the life sciences with 20+ years providing consulting and training services to pharmaceutical and medical device companies and other life science enterprises. Archives
December 2025
Categories |
RSS Feed