MCCULLEY/CUPPAN INC.
  • Home
  • Books
  • Consulting
    • Strategic Review
    • Assessment Services
  • Training
    • Document Standards
    • Skills Development Workshops
  • About
    • Experience
    • Client List
    • Blog
  • Contact

eCTD v4.0 More Than a Format Change: A Shift Toward Digital, Decision-ReadySubmissions

9/28/2025

0 Comments

 
For years, eCTD v3.2.2 has been the backbone of global regulatory submissions. It gave structure,
consistency, and a common language for sponsors and regulators. But today, our products and the data
behind them have outgrown the rigid folder hierarchy of this eCTD platform. eCTD 4.0 is not just a
technical upgrade. It is a shift toward data-centric, decision-ready submissions designed to match how
regulators actually read and decide.

Why eCTD 4.0—Built for modern product complexity
For eCTD v3.2.2, the backbone of a submission is a fixed folder tree. Each document must be placed in a rigid location with a pre-defined name. A concept that works well for small molecules and many biologic submissions with predictable sections. But this rigidity creates friction when you add new data streams. For example, emerging therapies like CAR-T cell products and CRISPR-based gene editing make this challenge sharper. These products generate new evidence categories. Such as, chain-of-identity and chain-of-custody records for each patient’s cells, highly specialized potency assays, and genomic off- target analysis for edited cells. Under eCTD v3.2.2, sponsors often wedge these data sets into ill-fitting sections or invent new folder labels

The result is added complexity for the regulatory reviewer. Instead of following an established mental map, regulators must pause to interpret where the information sits, why it’s there, and potentially how it links to other evidence. That extra interpretive work most likely increases cognitive load—the mental effort needed to find and integrate meaning. As I have discussed in other articles, high cognitive load slows review, raises the risk of misinterpretation, and invites clarifying questions or requests for additional information.

eCTD 4.0 reduces these problems with a metadata-first model. Instead of forcing content into a physical place, each document or data set is described by rich metadata. The metadata provides terms for data type, data source, and regulatory intent. These metadata elements become the “coordinates” for regulators to find what they need.

On top of metadata sits “Context-of-Use” tagging. This is a way to declare why the document exists and how it is meant to be used in the dossier review. For example, a potency assay file might carry context tags such as:
  • Supports product comparability across manufacturing sites
  • Establishes control strategy for viral vector potency
This context is an advance organizer that tells reviewers the intended purpose of evidence without requiring a narrative workaround.

In information design, as I teach it in my workshops, an advance organizer is a signal placed before complex material to help the reader build a mental map. By explicitly stating why a document or data set exists and how it will be used in decision-making enables Context-of-Use tags to work in the same way. The tags prepare reviewers to integrate what they are about to read, reduce search time, and anchor interpretation. Instead of scrolling between sections or inferring meaning from file names, the reviewer sees at a glance the analytical or regulatory task the content supports.

This is a cognitive service.

Over the course of a 200,000+ page submission, reducing even small interpretive friction compounds into meaningful efficiency and fewer avoidable queries.

The result: a scalable submission structure. You can introduce new modalities and data types without breaking the CTD map. Reviewers can filter and navigate by function and relevance, not just by where you tucked the file. And you can maintain submission integrity over time — updates to metadata or context can clarify meaning without re-life cycling the entire file.

Instead of duplicating entire documents across IND, NDA, BLA, and global variations, eCTD 4.0 uses unique document identifiers and machine-readable metadata to let the same content unit be updated and referenced throughout a product’s lifecycle. This reduces redundant authoring, helps maintain consistency across submissions, and supports automation for publishing and review.

Lastly, richer metadata, stable document identity is now critical as regulatory agencies pilot AI tools for comparison, safety signal detection, and labeling review. A well-tagged eCTD 4.0 dossier will be easier for both humans and machines to navigate.

Global Status
  • FDA began accepting eCTD 4.0 submissions in September 2024.
  • Japan is piloting and aiming for mandatory use in 2026.
  • The EU and other regulators are preparing pilots.
  • A mixed-mode period is coming — many sponsors will maintain both v3.2.2 and v4.0 workflows for the next several years.
The Opportunity
For organizations ready to adapt, eCTD 4.0 isn’t just compliance work. It’s a chance to modernize regulatory writing, improve clarity and traceability, and align better with how reviewers — and soon their AI assistants — read and analyze submissions.

Further Reading
  • FDA & ICH — eCTD v4.0 Implementation Guide
    Official specification covering metadata-first structure, submission-unit messaging, and context-of-use tagging. Download PDF
  • EMA eSubmission — Business Cases & Advantages of eCTD 4.0
    European perspective on business drivers, reuse, and harmonization of metadata-driven submissions. Download slides
  • Veeva — Plan for Submission Success with eCTD 4.0
    Practical guidance on context-of-use, document reuse, and preparing publishing systems and teams. Read here
0 Comments

Designing Document Links for Humans and Machines: Getting It Right

9/22/2025

0 Comments

 
So what? The way regulatory and medical writers design document hyperlinks now influences not only how efficiently human reviewers read but also how effectively AI tools, like FDA’s Elsa, parse, cross- reference, and retrieve content. Link design is no longer a simple way to manage information economy as defined within lean writing precepts. Link design directly shapes comprehension for humans and interpretability for machines.
​
The Human Reader Challenge
Many biopharma document users must read under time pressure. They work within and across long documents with dense appendices, moving between summary sections, data, methods, and supporting detail. Poorly structured links force them to jump back and forth with little context. That jump is not harmless.

Cognitive psychology highlights two distinct challenges:
  • Split attention effect. Readers must hold partial, spatially separated information in working memory while navigating elsewhere for the missing pieces. The extra load strains memory capacity, leaving less mental energy for comprehension.
  • Misdirected attention effect. In-text hyperlinks require constant micro-decisions about whether to click (what I refer to as the “should I stay or go now” phenomenon), increasing cognitive load and diverting attention from the main text. Poorly conceived links send readers into irrelevant or low-value detail. When they return, context is lost and momentum broken. Attention has been wasted on material that does not advance the argument.
For example:
  • A phrase such as “see appendix for details” without previewing what is in that appendix increases cognitive load.
  • A chain of multiple links (from main text → appendix → sub-appendix → external document) fragments attention further.
  • A sentence that presents three or more hyperlinks in sequence—such as “refer to 5.3.2, 6.4.3.1, 8.4.5.3”—forces readers to scatter attention across several targets at once, with no clear priority. The reader must also decide: Do I need to check all three? In what order? After reviewing one section, do I return to the original sentence before moving to the next? Each of these decisions adds unnecessary cognitive load and fragments comprehension.
Link formatting is not neutral—it influences what stands out and what gets lost. The Von Restorff effect shows that distinctive items are more likely to be noticed and remembered. Purposeful formatting of critical links can therefore guide reviewers’ attention to essential evidence. But overuse or inconsistent styling dilutes the effect, turning the page into visual noise. Instead of helping reviewers prioritize what matters, the links compete for attention and distract from the argument.

The result: reviewers not only lose the argument thread but also misinterpret the evidence when links overshoot or underserve their purpose. Poor link design magnifies two distinct risks—split attention and misdirected attention—each undermining comprehension in different ways.

The AI Reader Challenge
Hyperlinks can both improve and challenge Natural Language Processing (NLP) parsing in large technical documents by providing additional context and structure while also introducing noise and formatting complexities. For modern Large Language Models (LLMs), hyperlinks can be invaluable resources to enhance the quality of text-based applications like Retrieval-Augmented Generation (RAG).

AI tools such as FDA’s Elsa approach links differently. Machines do not skim, infer, or guess. They parse hierarchies and rely on structure. A vague cross-reference like “see above” or “refer to Appendix 1” leaves a machine with no anchor point.
For AI:
  • Consistency matters. A link must follow a stable format—such as “Appendix 3.2.S.4.1 Dissolution Data”—not shorthand like “see Table in Appendix.”
  • Persistence matters. Anchors must point to stable IDs or tags, not text that may shift during drafting.
  • Context matters. Machines need metadata around the link to understand relationships. For example, is the link pointing to supporting evidence, regulatory precedent, or comparative data?

The anchor text of a hyperlink—the clickable words—often provides a concise, semantically meaningful summary of the linked content. An NLP model can use this information to better understand the linked document’s topic and relevance. Internal links that connect different sections of the same document can act as a roadmap, informing NLP models of the document’s structure, similar to how a table of contents functions.

But not all anchor text is descriptive. Hyperlinks with generic text like “refer to Section 6.4.2” or “see Appendix 1” introduce noise for NLP systems, as they provide no semantic information. Another factor is technical documents in various formats, such as PDFs or legacy file types, may lack consistently marked- up hyperlinks—posing a major challenge for accurate hyperlink extraction.

Without consistent structure and metadata, AI parsing has constrained value. Tools may mis-index or
mis-categorize evidence, leading to gaps in automated review or flawed analytics.

The Dual Design Challenge
The challenge for regulatory writers is designing links that serve two audiences at once.
Human-centered design:
  • Preview what the reader will find on the other side of the link.
  • Integrate the link into the sentence logically, so readers don’t lose context.
  • Minimize unnecessary toggling by including summaries or excerpts before the link.
Machine-centered design:
  • Use structured patterns—consistent numbering, explicit section identifiers.
  • Anchor to stable, persistent IDs rather than vague references.
  • Provide metadata or descriptive labels that help AI categorize the relationship.

These principles are complementary, not competing. Links designed well reduce cognitive load for
people and improve interpretability for machines.

Why This Matters Now
Two shifts make link design urgent:
  • Human workload is rising. Regulatory reviewers face expanding data volumes. Poor navigation multiplies their effort and increases the risk of oversight.
  • AI oversight is accelerating. Tools like Elsa are entering mainstream regulatory review. Documents that lack structured, machine-readable links may slow automated checks or create mistrust in sponsor submissions.
In short, link design is a risk management decision. If links distract the human or confuse the machine,
the regulatory argument weakens.

Design Principles Going Forward
  • Preview before you link. Give readers context so they know why they are leaving the page. This reduces split attention by keeping meaning in view and prevents misdirected attention by signaling whether the link is relevant.
  • Reduce toggling. Summarize supporting evidence inline before directing to an appendix. This keeps working memory free (limiting split attention) and ensures only essential links are followed (limiting misdirected attention).
  • Think metadata. Add descriptive tags or labels clarifying the link’s purpose—whether it supports evidence, methods, or regulatory precedent. This helps AI parse relationships and signals to human readers whether the detail is worth following.
  • Audit your links. Review for vague references, irrelevant detail, and inconsistent styles. Each unchecked problem adds either memory strain (split attention) or wasted effort (misdirected attention).
Bottom Line
Future-ready regulatory documents must support two modes of reading: fast, context-seeking human
review and precise, structure-dependent AI parsing. Writers who treat links as part of information
design—not just formatting—reduce cognitive load for reviewers today and build trust with AI systems
tomorrow.

The real question is not whether your documents contain links. The real question is: Are your links
designed for humans and machine?

https://www.linkedin.com/pulse/designing-document-links-humans-machines-getting-right-gregory-
cuppan-veeec
0 Comments

Why Readability Is the Hidden Currency of Medical and Regulatory Writing

9/2/2025

0 Comments

 
​Readability determines whether your document works—or fails—in practice. A protocol, briefing book,
or other regulatory submission document may be perfectly accurate, but if readers struggle to navigate
the document, they most likely waste time, may make mistakes, and likely lose confidence in the work.

Working definition: readability is the degree to which a document enables the intended readers to
quickly find, understand, and apply the information with minimal cognitive effort.

In regulatory submission documents and protocols, readability is not optional—it is risk mitigation. High readability:
  • Reduces cognitive load for busy reviewers.
  • Prevents operational errors in study conduct.
  • Speeds decision-making by presenting information logically and without clutter.
When readability breaks down, interpretive space grows, and that gap between what’s written and
what’s understood becomes dangerous.

The Hidden Cost of Poor Readability

Poorly designed documents:
  • Inflate reading timelines.
  • Trigger avoidable questions.
  • Increase site errors in clinical research protocols.
These costs often remain invisible until after the document is published and is use—when problems are
hardest to fix.

Readability in technical and regulatory documents is not a cosmetic feature—it is a competitive advantage. As Saul Carliner observed, this is the 2nd level of information design: enabling documents to perform reliably in real-world use. He also suggests that when your documents reduce cognitive strain, you build trust with readers.

The most successful submission documents I’ve reviewed are not only scientifically rigorous, they are
designed to be read.

The Writer’s Responsibility
Patricia Wright stresses the author’s role in designing readable documents: “The message is that the onus for achieving successful communication cannot be safely left to the reader. Writers need to see themselves as catalysts for the strategies that their readers adopt; and they need to be aware of the design features that promote the selection of particular strategies.”

Wright’s insight shifts accountability squarely onto the author’s shoulders. Too often, medical and regulatory writers assume that expert readers will “figure it out” even if a passage is dense or disorganized. That assumption is dangerous in regulatory contexts, where readers work under time constraints, juggle multiple documents, and must reach reliable conclusions.

​Readable writing is not about lowering standards—it is about following document design standards as
an act of responsibility. Authors shape the strategies readers use and they create conditions that
promote consistent, accurate, and rapid comprehension.
0 Comments

The Hidden Cognitive Cost of Hyperlinks in High-Stakes Documents

9/2/2025

0 Comments

 
​Hyperlinks are everywhere—in protocols, briefing books, submission documents, SOPs, policy manuals, and training guides. Their siren song promises speed, efficiency, and instant navigation.

In Greek mythology, the Sirens’ enchanting melodies lured sailors off course and onto rocky shores. Hyperlinks can work the same way: they invite you to click, to leave the safe harbor of your main discussion or argument in search of something interesting, only to risk losing your place, your context, and sometimes the point entirely.

In clinical research protocols, they may link eligibility criteria to lab thresholds, dosing schedules to product handling instructions, or definitions to appendices. For AI and other automated tools, these links are gold: they can map relationships between sections, create a machine-readable network of content, and cross-check for consistency.

For human readers, however, hyperlinks come with a trade-off: every link is a micro-decision. When you see a hyperlink, your brain has to ask:
  • Should I stay or should I go now? (to borrow a line from the music group The Clash)
  • Will I lose my place if I follow it?
  • Will it answer my question, or lead me somewhere irrelevant?
That momentary pause—repeated dozens of times in a dense protocol or Module 2 summary—splits attention, interrupts flow, and increases the risk of losing context. Cognitive science calls this the split attention effect. This is when a reader must divide focus between two information sources, the effort of mentally integrating them increases cognitive load and reduces comprehension. In high-stakes documents, that extra mental friction can mean slower decisions, missed connections, or misinterpretation of the intended message.

In operational documents like clinical trial protocols, GxP SOPs, or emergency response manuals, these interruptions are more than an inconvenience. They can delay decisions, increase errors, and erode compliance.

In regulatory briefing books and Module 2 documents, hyperlink interruptions can undermine both efficiency and precision. Each click risks pulling reviewers away from the main argument, breaking the logical chain that supports a decision. Re-orienting after navigating to annexes, study reports, or external references slows the evaluation process and can erode the clarity of your case. Strategic hyperlinking should serve the narrative—pointing to critical evidence only when it truly strengthens comprehension—rather than scattering attention across disconnected content.

Another trap in protocol and submission writing is embedding full source links directly into the main
narrative, as if the link itself proves transparency or credibility. In practice, this type of linkage clutters
the discussion, distracts the reader, and breaks the flow. Demonstrating a link to source is
essential—but the primary text should focus on instruction, reasoning, interpretation, or conclusion.

Make the appearance of secondary and tertiary links as subordinate as their intention. Use super script notation to link to footnotes, references, or appendices. The goal is to make these supportive connections accessible without becoming visual speed bumps for your reader.

Like the sailors of myth, regulatory and medical writers must recognize when the Sirens are singing. Hyperlinks can be useful guides, but if they tempt the reader away from the main course of the discussion or argument, the hyperlink risks wrecking clarity on the rocky shores of distraction. The safest passage is not to silence the Sirens, but to decide when their song strengthens the voyage—and when it should be left unheard.
0 Comments

    Author

    Gregory Cuppan is the Managing Principal of McCulley/Cuppan Inc., a group he co-founded. Mr. Cuppan has spent 30+ years working in the life sciences with 20+ years providing consulting and training services to pharmaceutical and medical device companies and other life science enterprises.

    View my profile on LinkedIn

    Archives

    December 2025
    November 2025
    October 2025
    September 2025
    July 2025
    June 2025
    April 2025
    August 2019
    April 2019
    March 2019
    May 2016

    Categories

    All

    RSS Feed

Services

Consulting
Training
Assessment

Company

About
Experience
Blog

Support

Contact
© COPYRIGHT 2022. ALL RIGHTS RESERVED.
Picture
  • Home
  • Books
  • Consulting
    • Strategic Review
    • Assessment Services
  • Training
    • Document Standards
    • Skills Development Workshops
  • About
    • Experience
    • Client List
    • Blog
  • Contact