Accelerating Technology Adoption: Reflections and Predictions

 Listen to Audio

 

Listen Anywhere

December 2025

In this episode, host Bob Chew shares insights on accelerating technology adoption in the pharmaceutical industry.  Dive into this discussion to explore the future of pharma innovation, including transitioning from document-centric to data-driven quality systems, leveraging AI and digital twins for enhanced manufacturing processes, and reimagining risk management with statistical tools.

  • Guest

    Robert Chew
    Chair of the Board of Directors
    CAI
  • Transcript

    Back to Top

    Download Transcripts


     

    1

    00:00:00,000 --> 00:00:05,000

    Welcome to the ISPE podcast, shaping the future of pharma, where ISPE

    2

    00:00:05,000 --> 00:00:10,000

    supports you on your journey, fueling innovation, sharing insights, thought leadership,

    3

    00:00:10,000 --> 00:00:14,000

    and empowering a global community to reimagine what's possible.

    4

    00:00:15,000 --> 00:00:20,000

    Hello, and welcome to the ISPE podcast, Shaping the Future of Pharma.

    5

    00:00:21,000 --> 00:00:22,000

    I'm Bob Chew, your host.

    6

    00:00:23,000 --> 00:00:29,000

    And today, we have another episode where we'll be sharing the latest insights

    7

    00:00:27,000 --> 00:00:32,000

    and thought leadership on manufacturing, technology, supply chains

    8

    00:00:32,000 --> 00:00:36,000

    and regulatory trends impacting the pharmaceutical industry.

    9

    00:00:37,000 --> 00:00:42,000

    You will hear directly from the innovators, experts and professionals driving progress

    10

    00:00:42,000 --> 00:00:44,000

    and shaping the future.

    11

    00:00:45,000 --> 00:00:49,000

    Thank you again for joining us, and now let's dive into this episode.

    12

    00:00:50,000 --> 00:00:54,000

    Our topic today is accelerating technology adoption.

    13

    00:00:55,000 --> 00:01:00,000

    I am recording this podcast from a National Geographic ship in the Antarctic.

    14

    00:01:01,000 --> 00:01:06,000

    If you hear occasional crashes or sounds like thunder, that is likely the

    15

    00:01:06,000 --> 00:01:08,000

    ship moving through the ice pack.

    16

    00:01:09,000 --> 00:01:15,000

    I have no special guest for this episode and will instead offer my thinking on this

    17

    00:01:15,000 --> 00:01:18,000

    topic of accelerating technology adoption.

    18

    00:01:19,000 --> 00:01:24,000

    As part of doing this podcast series, I attend many ISPE

    19

    00:01:24,000 --> 00:01:30,000

    and other industry conferences, hearing case studies, listening to executives,

    20

    00:01:30,000 --> 00:01:35,000

    regulators, and technical experts talk about the future and the present.

    21

    00:01:36,000 --> 00:01:41,000

    I hope you will find this thought provoking and perhaps create your own innovative

    22

    00:01:41,000 --> 00:01:45,000

    approaches to shape the future of pharmaceutical manufacturing.

    23

    00:01:47,000 --> 00:01:52,000

    In this session, I'm going to discuss three broad topics: documents, data,

    24

    00:01:53,000 --> 00:01:54,000

    and statistics.

    25

    00:01:55,000 --> 00:02:00,000

    Now what do these have to do with the subject of this podcast, Accelerating Technology

    26

    00:02:00,000 --> 00:02:01,000

    Adoption?

    27

    00:02:02,000 --> 00:02:08,000

    Well, innovative technologies such as AI and digital

    28

    00:02:08,000 --> 00:02:13,000

    twins are based on sophisticated statistical algorithms.

    29

    00:02:13,000 --> 00:02:19,000

    When we train these models, we feed them data, lots of data, data

    30

    00:02:19,000 --> 00:02:22,000

    that is relevant to what we want the model to do.

    31

    00:02:22,000 --> 00:02:27,000

    The more data, the more statistically relevant outputs we will get.

    32

    00:02:28,000 --> 00:02:33,000

    Of course, large language models are trained based on language, I e,

    33

    00:02:34,000 --> 00:02:35,000

    documents.

    34

    00:02:36,000 --> 00:02:41,000

    Now GMP regulations contain many references to

    35

    00:02:41,000 --> 00:02:45,000

    and expectations of documents.

    36

    00:02:45,000 --> 00:02:49,000

    Our quality systems are largely document based.

    37

    00:02:50,000 --> 00:02:55,000

    We have policies, procedures, protocols, work instructions

    38

    00:02:56,000 --> 00:02:57,000

    by the thousands.

    39

    00:02:59,000 --> 00:03:00,000

    Let me ask this question.

    40

    00:03:01,000 --> 00:03:06,000

    Should our quality systems continue to be document based or should we

    41

    00:03:06,000 --> 00:03:09,000

    move towards a more data driven quality system?

    42

    00:03:11,000 --> 00:03:17,000

    Now, our production and process control systems provide data

    43

    00:03:17,000 --> 00:03:20,000

    by the boatload continuously.

    44

    00:03:21,000 --> 00:03:25,000

    Separately, we have data generated by our QC laboratories.

    45

    00:03:26,000 --> 00:03:32,000

    Separately, we have information about human performance, including training records.

    46

    00:03:33,000 --> 00:03:38,000

    Today, technology exists to monitor operators and assess human performance,

    47

    00:03:39,000 --> 00:03:44,000

    which could focus on how the operators contribute to process and contamination control

    48

    00:03:45,000 --> 00:03:50,000

    and record proper or improper aseptic techniques, for example.

    49

    00:03:50,000 --> 00:03:56,000

    Thus, we can even include information about human performance

    50

    00:03:56,000 --> 00:03:59,000

    into the category of data.

    51

    00:04:01,000 --> 00:04:06,000

    Now, new technologies and computing power exist to synthesize

    52

    00:04:06,000 --> 00:04:11,000

    all the above and create a production model which could

    53

    00:04:11,000 --> 00:04:16,000

    analyze the effectiveness of the process and contamination control strategies

    54

    00:04:17,000 --> 00:04:22,000

    and even identify opportunities for improvement through statistical evaluation.

    55

    00:04:24,000 --> 00:04:29,000

    In other words, data, a data driven quality system.

    56

    00:04:31,000 --> 00:04:36,000

    Let's examine how data could supplant documents as the foundation of

    57

    00:04:36,000 --> 00:04:37,000

    our quality system.

    58

    00:04:38,000 --> 00:04:43,000

    Today, when we have a deviation, we create a document.

    59

    00:04:44,000 --> 00:04:49,000

    That document describes what happened and offers root cause analysis,

    60

    00:04:50,000 --> 00:04:53,000

    corrective actions, and preventive actions for the future.

    61

    00:04:54,000 --> 00:04:59,000

    We agonize over the wording of these documents since they will be

    62

    00:04:59,000 --> 00:05:04,000

    used to explain to future inspectors what really happened and

    63

    00:05:04,000 --> 00:05:09,000

    provide a convincing case that the corrective and preventive actions were

    64

    00:05:09,000 --> 00:05:12,000

    both justified and sufficient.

    65

    00:05:14,000 --> 00:05:20,000

    What if in the future we had a combination of Process Digital Twin constantly

    66

    00:05:20,000 --> 00:05:25,000

    monitoring the process and an AI platform that integrated

    67

    00:05:25,000 --> 00:05:31,000

    past deviation data, past what happened, what

    68

    00:05:31,000 --> 00:05:35,000

    were the corrective actions, and what were the results of those actions.

    69

    00:05:36,000 --> 00:05:42,000

    When we have the next deviation, these technologies can provide a statistically

    70

    00:05:42,000 --> 00:05:47,000

    ranked set of possible root causes with associated

    71

    00:05:47,000 --> 00:05:53,000

    probabilities and recommended corrective actions that are based on observed

    72

    00:05:53,000 --> 00:05:56,000

    effectiveness of similar actions in the past.

    73

    00:05:57,000 --> 00:06:02,000

    The human can still provide the machine generated still approve the

    74

    00:06:02,000 --> 00:06:07,000

    machine generated recommendation but using statistics based on data.

    75

    00:06:08,000 --> 00:06:14,000

    And from time to time, the human may identify new

    76

    00:06:14,000 --> 00:06:19,000

    root causes and new opportunities for innovative improvements.

    77

    00:06:20,000 --> 00:06:24,000

    But we no longer agonize over the wording of documents.

    78

    00:06:27,000 --> 00:06:32,000

    Further, today, our conclusions can often be operator error

    79

    00:06:32,000 --> 00:06:35,000

    or laboratory error.

    80

    00:06:35,000 --> 00:06:41,000

    If AI is assessing these situations based on data, it

    81

    00:06:41,000 --> 00:06:46,000

    might point to other root causes based on the real data and the statistical

    82

    00:06:46,000 --> 00:06:51,000

    probability that it really isn't operator error or

    83

    00:06:51,000 --> 00:06:53,000

    laboratory error.

    84

    00:06:54,000 --> 00:06:59,000

    Let's broaden the conversation and consider quality risk management.

    85

    00:07:00,000 --> 00:07:05,000

    First, I'd like to remind ourselves that each and every GMP

    86

    00:07:05,000 --> 00:07:11,000

    regulation is designed to control a risk to

    87

    00:07:11,000 --> 00:07:12,000

    manufacturing quality.

    88

    00:07:13,000 --> 00:07:18,000

    In other words, the original quality risk management system is our

    89

    00:07:18,000 --> 00:07:19,000

    GMP regulations.

    90

    00:07:20,000 --> 00:07:25,000

    For example, GMP regulations tell us that we must have procedures

    91

    00:07:25,000 --> 00:07:30,000

    for cleaning and that those cleaning procedures

    92

    00:07:30,000 --> 00:07:32,000

    be validated.

    93

    00:07:32,000 --> 00:07:34,000

    Why is that?

    94

    00:07:34,000 --> 00:07:39,000

    Well, it is to control the risk that cleaning is

    95

    00:07:39,000 --> 00:07:44,000

    not sufficient and that it doesn't clean properly.

    96

    00:07:45,000 --> 00:07:48,000

    And so we have these regulations.

    97

    00:07:49,000 --> 00:07:54,000

    How might we use a statistically based technology to automatically

    98

    00:07:54,000 --> 00:07:56,000

    generate risk assessments?

    99

    00:07:57,000 --> 00:08:02,000

    Well, if we had an AI model of aseptic

    100

    00:08:02,000 --> 00:08:08,000

    processing And that AI model

    101

    00:08:08,000 --> 00:08:13,000

    was trained on a combination of regulations

    102

    00:08:13,000 --> 00:08:18,000

    for aseptic process control like Annex one.

    103

    00:08:18,000 --> 00:08:23,000

    If it was trained on how equipment is designed

    104

    00:08:23,000 --> 00:08:29,000

    for aseptic, control, contamination control, if it was

    105

    00:08:29,000 --> 00:08:34,000

    trained on the actual observed effectiveness of those controls

    106

    00:08:34,000 --> 00:08:39,000

    during ongoing aseptic production operations, we

    107

    00:08:39,000 --> 00:08:45,000

    could then have a tool that is automatically updated around

    108

    00:08:45,000 --> 00:08:48,000

    the concept of contamination control.

    109

    00:08:51,000 --> 00:08:56,000

    And it would be assessing risks, new risks,

    110

    00:08:56,000 --> 00:09:01,000

    old risks, how effective are those controls, and constantly

    111

    00:09:01,000 --> 00:09:06,000

    reporting out that these controls are effective and these controls

    112

    00:09:06,000 --> 00:09:09,000

    might need, improvements.

    113

    00:09:11,000 --> 00:09:16,000

    I think it would be super powerful if we could apply, statistical

    114

    00:09:16,000 --> 00:09:22,000

    techniques through process digital twins, through aseptic

    115

    00:09:22,000 --> 00:09:27,000

    process digital twins to generate automatically

    116

    00:09:27,000 --> 00:09:29,000

    quality risk management metrics.

    117

    00:09:32,000 --> 00:09:37,000

    Now how might equipment digital twins be used to accelerate startup

    118

    00:09:37,000 --> 00:09:39,000

    and qualification of equipment?

    119

    00:09:40,000 --> 00:09:45,000

    After all, digital twins of jet engines have been used for a long time to

    120

    00:09:45,000 --> 00:09:50,000

    analyze performance and to model failure modes and effects.

    121

    00:09:51,000 --> 00:09:54,000

    Well, imagine an aseptic filling line.

    122

    00:09:55,000 --> 00:10:01,000

    Imagine a digital twin of such a filling line, which, by the way, has

    123

    00:10:01,000 --> 00:10:06,000

    already been created and used in at least one aseptic manufacturing

    124

    00:10:06,000 --> 00:10:08,000

    plant in Ireland.

    125

    00:10:08,000 --> 00:10:13,000

    There was a case study presented at an ISPE conference about a year ago.

    126

    00:10:15,000 --> 00:10:19,000

    Imagine training this digital twin during factory acceptance testing.

    127

    00:10:21,000 --> 00:10:26,000

    And so now this digital twin has a representation,

    128

    00:10:26,000 --> 00:10:31,000

    a model of how that aseptic line should work, how each and every

    129

    00:10:31,000 --> 00:10:37,000

    manipulation on that line should work, its performance.

    130

    00:10:37,000 --> 00:10:43,000

    Now we disassemble the line, we ship it, we set it up at the, manufacturing

    131

    00:10:43,000 --> 00:10:48,000

    site, and as we go through the startup process, we plug

    132

    00:10:48,000 --> 00:10:50,000

    in the digital twin.

    133

    00:10:51,000 --> 00:10:56,000

    It monitors how the machine is operating, and it's able to point

    134

    00:10:56,000 --> 00:11:01,000

    that you need to adjust, make this adjustment on this part.

    135

    00:11:01,000 --> 00:11:04,000

    You need to look at this connection.

    136

    00:11:04,000 --> 00:11:05,000

    It's not working right.

    137

    00:11:06,000 --> 00:11:11,000

    But at the end of the day, it's able to automatically say

    138

    00:11:11,000 --> 00:11:14,000

    that the equipment is qualified.

    139

    00:11:15,000 --> 00:11:16,000

    Wouldn't that be sweet?

    140

    00:11:17,000 --> 00:11:23,000

    Today, AI writes computer code and

    141

    00:11:23,000 --> 00:11:26,000

    it self tests that computer code.

    142

    00:11:27,000 --> 00:11:32,000

    Not necessarily in the pharmaceutical industry, but that technology exists.

    143

    00:11:32,000 --> 00:11:37,000

    Imagine being able to apply that technology to automatically generate

    144

    00:11:37,000 --> 00:11:43,000

    the process control automation software and

    145

    00:11:43,000 --> 00:11:45,000

    to automatically test it.

    146

    00:11:47,000 --> 00:11:52,000

    That would certainly be an acceleration of the delivery of manufacturing capability.

    147

    00:11:54,000 --> 00:11:59,000

    Now I'm going to suggest a break from past approaches to process validation.

    148

    00:12:01,000 --> 00:12:07,000

    First, unless you're doing a site to site transfer, it's

    149

    00:12:06,000 --> 00:12:12,000

    unreasonable to expect that development batches supplemented

    150

    00:12:12,000 --> 00:12:17,000

    with a few engineering batches will yield a statistically valid

    151

    00:12:17,000 --> 00:12:18,000

    design space.

    152

    00:12:19,000 --> 00:12:20,000

    Let me say that again.

    153

    00:12:21,000 --> 00:12:26,000

    Design space is that three-dimensional or multi dimensional

    154

    00:12:26,000 --> 00:12:32,000

    model of where you can manufacture and get a

    155

    00:12:32,000 --> 00:12:33,000

    successful batch.

    156

    00:12:35,000 --> 00:12:39,000

    That takes quite a few statistical, batches to build.

    157

    00:12:41,000 --> 00:12:46,000

    Nor can we really say that a process is truly validated based on three successful

    158

    00:12:46,000 --> 00:12:52,000

    batches, which are not always without a failed batch or two interspersed.

    159

    00:12:54,000 --> 00:12:59,000

    I will offer that a robust process digital twin

    160

    00:12:59,000 --> 00:13:05,000

    is nothing more than the design space from an underlying statistics

    161

    00:13:05,000 --> 00:13:06,000

    perspective.

    162

    00:13:07,000 --> 00:13:11,000

    It takes quite a few batches to train a digital twin.

    163

    00:13:12,000 --> 00:13:17,000

    But once you have it trained, you also have a robust

    164

    00:13:17,000 --> 00:13:20,000

    design space defined.

    165

    00:13:21,000 --> 00:13:26,000

    As we then continue to manufacture and problems arise, the digital

    166

    00:13:26,000 --> 00:13:31,000

    twin can either point to the root cause or

    167

    00:13:31,000 --> 00:13:36,000

    it can say unknown new source of variation identified,

    168

    00:13:37,000 --> 00:13:42,000

    which might be that you change bioreactor bag supplier or the supplier altered

    169

    00:13:42,000 --> 00:13:47,000

    its methods or raw materials without understanding the impact on your process.

    170

    00:13:49,000 --> 00:13:54,000

    In short, getting a process to a validated

    171

    00:13:54,000 --> 00:13:59,000

    state is a continuous process and it will take

    172

    00:13:59,000 --> 00:14:05,000

    quite a few batches to really develop that robust design space.

    173

    00:14:06,000 --> 00:14:11,000

    Okay, you might agree with me theoretically, but from a practical perspective,

    174

    00:14:12,000 --> 00:14:16,000

    when would regulators give the green light to market the product?

    175

    00:14:16,000 --> 00:14:19,000

    How many batches must be produced first?

    176

    00:14:20,000 --> 00:14:25,000

    Well, I can imagine a regulator having an AI tool

    177

    00:14:26,000 --> 00:14:31,000

    that statistically assesses the degree of control exhibited by the

    178

    00:14:31,000 --> 00:14:37,000

    process as reflected in consistent process performance data.

    179

    00:14:38,000 --> 00:14:43,000

    The licensed applicant would have the same data and would notify regulators

    180

    00:14:43,000 --> 00:14:48,000

    when it feels that it has its process in a full state of control, regardless

    181

    00:14:48,000 --> 00:14:50,000

    of how many batches that is.

    182

    00:14:51,000 --> 00:14:56,000

    The regulator would access the process data and,

    183

    00:14:56,000 --> 00:15:02,000

    using its AI tool, analyze the data and render an approval decision.

    184

    00:15:03,000 --> 00:15:06,000

    This would automate and accelerate the licensing process.

    185

    00:15:07,000 --> 00:15:12,000

    Couple this with machine digital twins reporting out to regulators that a qualified

    186

    00:15:12,000 --> 00:15:17,000

    state has been attained, and we take the guesswork and time delays

    187

    00:15:18,000 --> 00:15:20,000

    out of, preapproval inspections.

    188

    00:15:24,000 --> 00:15:29,000

    If we agree that new technologies based on statistics will drive us

    189

    00:15:29,000 --> 00:15:34,000

    towards a data centric quality system, then data integrity

    190

    00:15:34,000 --> 00:15:38,000

    takes on even more importance than it has up to now.

    191

    00:15:39,000 --> 00:15:44,000

    Data integrity has focused on ensuring that data meets the ALCOA

    192

    00:15:44,000 --> 00:15:50,000

    plus characteristics of attributable, legible, contemporaneous,

    193

    00:15:50,000 --> 00:15:56,000

    original, accurate, complete, consistent, enduring,

    194

    00:15:56,000 --> 00:15:57,000

    and available.

    195

    00:15:58,000 --> 00:16:04,000

    But in addition, when it comes to training an AI or digital twin model, the

    196

    00:16:04,000 --> 00:16:09,000

    data must also be relevant to the situation or application being

    197

    00:16:09,000 --> 00:16:10,000

    trained.

    198

    00:16:10,000 --> 00:16:15,000

    Now I will leave it to our Pharma four point o colleagues to expound upon this dimension

    199

    00:16:15,000 --> 00:16:18,000

    of data and data integrity.

    200

    00:16:19,000 --> 00:16:24,000

    So back to the title of this podcast episode, Accelerating Technology

    201

    00:16:24,000 --> 00:16:25,000

    Adoption.

    202

    00:16:26,000 --> 00:16:28,000

    How can we achieve this acceleration?

    203

    00:16:29,000 --> 00:16:34,000

    My answer is, by recognizing how these new technologies really

    204

    00:16:34,000 --> 00:16:39,000

    work, we accept the use of statistics to assess clinical outcomes,

    205

    00:16:39,000 --> 00:16:41,000

    the basis for all drug approvals.

    206

    00:16:42,000 --> 00:16:48,000

    We understand that quality risk management is founded on probabilistic estimates

    207

    00:16:48,000 --> 00:16:50,000

    of hazards and controls.

    208

    00:16:51,000 --> 00:16:56,000

    Therefore, we should embrace the power of new statistical tools of

    209

    00:16:56,000 --> 00:17:02,000

    which AI and digital twins are prime examples and implement

    210

    00:17:02,000 --> 00:17:05,000

    such tools across the manufacturing operation.

    211

    00:17:06,000 --> 00:17:12,000

    In conjunction, we need to deemphasize documents and base

    212

    00:17:12,000 --> 00:17:14,000

    our quality systems on data.

    213

    00:17:15,000 --> 00:17:20,000

    Both industry and regulators should come together to discuss how

    214

    00:17:20,000 --> 00:17:25,000

    to transform our industry to a nimble, data driven approach to process control,

    215

    00:17:26,000 --> 00:17:29,000

    contamination control, and quality assurance.

    216

    00:17:30,000 --> 00:17:36,000

    Old expectations of documents should be replaced by new expectations

    217

    00:17:36,000 --> 00:17:37,000

    of data analysis.

    218

    00:17:39,000 --> 00:17:45,000

    The quality function should evolve from one of compliance to procedures

    219

    00:17:45,000 --> 00:17:50,000

    to one of a true quality engineer, driving innovation and changes for

    220

    00:17:50,000 --> 00:17:51,000

    improvement.

    221

    00:17:52,000 --> 00:17:57,000

    We must get comfortable with the use of AI and digital twins for

    222

    00:17:57,000 --> 00:18:02,000

    automated process control and continued refinement of the design space.

    223

    00:18:04,000 --> 00:18:09,000

    In summary, if we view AI and digital twins from the lens

    224

    00:18:09,000 --> 00:18:14,000

    of statistics and if we can appreciate that the more data we feed these

    225

    00:18:14,000 --> 00:18:19,000

    tools, the more useful and robust they become, we conclude

    226

    00:18:19,000 --> 00:18:24,000

    that we should move away from a document centric quality system and towards a data

    227

    00:18:24,000 --> 00:18:26,000

    centric quality system.

    228

    00:18:27,000 --> 00:18:32,000

    Quality management throughout the life cycle of a product and facility becomes

    229

    00:18:32,000 --> 00:18:35,000

    a continuous improvement journey.

    230

    00:18:36,000 --> 00:18:41,000

    Each improvement, each change, should be made based on the statistics of

    231

    00:18:41,000 --> 00:18:47,000

    real data and in most cases can be implemented either automatically

    232

    00:18:47,000 --> 00:18:49,000

    or without a change control package.

    233

    00:18:51,000 --> 00:18:56,000

    That brings us to the end of another episode of the ISPE podcast, Shaping

    234

    00:18:56,000 --> 00:18:58,000

    the Future of Pharma.

    235

    00:18:59,000 --> 00:19:05,000

    Please be sure to subscribe so you don't miss future conversations with the innovators,

    236

    00:19:05,000 --> 00:19:09,000

    experts, and change makers driving our industry forward.

    237

    00:19:10,000 --> 00:19:15,000

    On behalf of all of us at ISPE, thank you for listening, and we'll see

    238

    00:19:15,000 --> 00:19:20,000

    you next time as we continue to explore the ideas, trends, and

    239

    00:19:20,000 --> 00:19:23,000

    people shaping the future of pharma.

     

Listen to Past Episodes