The GAMP (Good Automated Manufacturing Practice) guide version 5 was released in March 2008 and one of the changes was that the classification of software was revised - again. This column will look at what the changes mean for the laboratory and whether all of these should be implemented.
The GAMP (Good Automated Manufacturing Practice) guide version 5 was released in March 2008 and one of the changes was that the classification of software was revised again. This column will look at what the changes mean for the laboratory and whether all of these should be implemented.
Version 5 of the Good Automated Manufacturing Practice (GAMP) guide (1) was released last year. This publication has been available since 1994, when version 1 was informally published in the UK, and since its inception it has always contained a classification of software. This is one of the best parts of the guide as it has an in-built risk assessment, as we shall see in this column. We will explore version 5 of the software classification and see what changes we need to make to ensure that it can be implemented practically and effectively in any laboratory.
However, before we continue much further I should also declare a vested interest: I have a love–hate relationship with the GAMP guide. I love the classification of software outlined in Appendix M4 and hate the life cycle V model. My rationale for this position is that versions 1–4 of the guide presented a single life cycle V model that was really only applicable to process equipment and manufacturing systems. It had very little to do with computerized systems, especially laboratory ones. Therefore, every validation was shoehorned into an inappropriate model because there was little thought and intelligence applied and the model followed blindly. For example, when a commercially available laboratory system was validated, functional and design specifications were written for virtually no gain but at a great cost in time and resources. The problem lay in the origins of the GAMP guide. The first version was written by a group of volunteers in the UK in the early 1990s as a mechanism to control suppliers of process equipment to the pharmaceutical industry, and this legacy survived through to version 4. However, the model does not make it into version 5 of the Guide, which is a shame; as mentioned above, the model is very good for process equipment.
However in GAMP version 5, I'm very pleased to say that the "one size fits all" approach has been replaced by a breath of fresh air with different life cycles depending on the classification of the software being implemented. The key message is that now a single size life cycle model does not fit all systems. Note that GAMP is a guide and you can deviate from it — all that is required is the application of thought and intelligence coupled with effective risk management that is well documented. OK, perhaps this is a step too far . . . .
As I mentioned earlier, the software categories in GAMP 5 have been revised (1). To appreciate the scope of these changes fully we need to look at the classification of software from GAMP 4 (2) and compare this with GAMP 5, as shown in Table I.
Table I: Comparison of software categories in GAMP 4 and GAMP 5
In the beginning, or at least in GAMP 4, there were five categories of software:
The constituents of each category are outlined in Table I; however, there was always a debate about some commercial software packages — were they category 3 or 4? Many spectroscopists would argue that an application should be classified as category 3 and not 4, as it should be less work to validate and evade the real classification. To help resolve this debate, in GAMP 5 the software categories have been revised and refined — most for the better and one for the worse. This is a natural evolution of this approach to software classification. So we now have the following four categories:
Refer to Table I as we discuss the changes in the software classification in more detail in the next section.
Before we go into a detailed discussion of the software categories, perhaps we should ask the question "Why bother to classify software?". What benefit does this software classification provide?
If you look at Table I there is a built-in risk assessment. The least risky and most widely available software is in category 1 (operating systems, databases, office software, and other widely available software). This is widely available software that can be used by anyone and in any industry. As we progress through down the categories as shown in Table I, generally the software becomes more specialized in its function (from a general office application to software that can control a spectrometer to acquire and process data then report the results). As we go down the list there is the increasing ability of the users to change the operation of software and process the results until we reach category 5. In category 5 is a unique solution that is conceived, specified, written, tested, and maintained by the users or the organization; here is the greatest risk. Let's now take a detailed look at each of the software categories and see what has changed and if there are any problems we need to discuss.
Presented and discussed here are the various changes to the software classifications in the new GAMP guide.
Category 1: Greatly Expanded Scope — Infrastructure Software
Category 1 has undergone a radical change and expansion from simply operating systems, that had been constant in GAMP versions 1 to 4, to infrastructure software. This category is broken down into two subcategories:
The intention is that the infrastructure software in this category provides the computing environment for running both regulated and nonregulated applications within an organization. All software in this category needs to be controlled and qualified in an organization so that dual standards are not applied by the IT department, which can question the status of validated applications if not done.
Software in the subcategory of Established or Commercially Available Layered Software still includes operating systems from GAMP 4, but this has also been expanded to encompass a greater scope:
The key issue is that many of these software tools are the base products for the applications used in the laboratory or they are the foundation layer for the laboratory applications to operate under. For example, if your spectrometer has application software that has a database to manage the methods, data, and results you generate, the latter is configured by the spectrometer supplier from the out-of-the-box database to operate with their application software. Languages are used as a means to write the applications, each of which will be validated for their intended use.
Note that category 1 also includes office software such as word processing, spreadsheet, database, and presentation applications. Now before you rush off thinking that Excel templates and macros do not need to be validated, think again, as the guide notes that "applications developed using these packages" are excluded from category 1 and these can be category 3, 4, or 5, respectively (1), depending on their complexity.
Note also the phrasing of the subcategory "established or commercially available". This means that both open source and software commercial can be used, which ratifies the status quo (open source operating systems [Linux or OpenVMS], databases [MySQL], and source code management [SubVersion]). In many IT departments and research groups open source software is used and often this use can be extensive. Some people may argue that open source software is hacked code, but when the code can be reviewed by many programmers it may be argued that the quality of the finished application could be better than some commercially available software. Regardless of the debate, the word established allows the use of open source applications within category 1.
The second subcategory is infrastructure software tools that comprise a wide variety of software, such as
While sounding like a shopping list, it provides the IT group with tools to establish, protect, monitor, and manage their computing environment and networks and where you store your laboratory data and electronic records. However, care needs to be taken with applications in this subcategory as the use of the application could drastically change the category to which it is allocated. Take for example help desk or configuration management applications. If no regulatory data are held exclusively here then they are category 1, but if they are your only tools for problem management or change control processes for regulated applications, then this changes the category and the validation group comes galloping round the bend.
From the laboratory, audit, and inspectional perspectives, what is required is control of the applications that comprise group 1. Some of the typical controls will be
All should be done regardless of where the software originates: open source or commercial software. Furthermore, change control and configuration management are essential elements of control both with category 1 and the other software categories. So we qualify these items of software, not validate them. In contrast, we validate software in categories 3, 4, and 5. Note that I have omitted category 2 software; we will now discuss this in more detail.
Category 2: Ignore Discontinuation of Firmware Classification — but with Care
As you can see in Table I, GAMP 4 had five categories of software, which is reduced to four in the latest version. The category that has gone missing in action is category 2 (firmware).
The argument for this category's discontinuation from the GAMP Forum is that firmware, which can vary from simple to custom software, can be accommodated in the other categories depending on its nature. To understand why this category was eliminated from GAMP 5 we need to consider what we mean by the term firmware. In its original form, firmware was a set of operating instructions for an instrument embedded in a read-only memory (ROM) chip or used to start more complicated programs on an instrument or device. An alternative term favored by IBM is microcode. This is software but instead of being delivered on a disk or USB stick, it comes preinstalled on a chip having been built into the instrument during the manufacturing process.
FOR: Firmware is still software, and different chip versions will be produced over time due to bugs that will be found and fixed by a manufacturer. When this occurs, a new version of the firmware ROM is produced. During a maintenance visit by a service engineer (aka the Angel of Death), the firmware ROM may be replaced with a new one. Therefore, change control must include firmware as the upgraded chip may contain new functions not present in the existing chip as well as the bug fixes for the original software. One instance where a firmware change may need to be upgraded is when you install a new version of instrument–data system application software. It is possible that the drivers for your spectrometer may have changed and to ensure that the new software works correctly, the firmware needs to be upgraded at the same time. You need to know that firmware will be upgraded in advance so you can complete the change request before the work starts.
Over time, firmware chips have become bigger to allow more instructions to be input and the ROM can be replaced with flash ROM, which can be erased by UV light and rewritten through a firmware updater. Other forms of firmware can be upgraded via a download; the best example of this is the BIOS firmware chip in your PC, which typically can be updated online to fix errors, improve functionality, and keep it current.
The biggest issue with firmware now is the ability to define programs or parametrization to produce user-defined routines that can be kept in memory and recalled at will. A typical example is a dispenser–dilutor, in which a user can prepare different routines for different analytical methods. So instead of a simple set of instructions there is now the ability of the users to change the way the instrument works. However, flexibility comes at a potential price of the impact on the results and errors if the user-defined program is incorrect. The parallel in the conventional software world is Excel and the ability to produce incorrect templates or macros if proper controls are not in place.
Therefore, the rationale for the discontinuation in GAMP 5 is that the software sitting on the firmware can be classified in the remaining software categories. In doing so it should eliminate the impact of user-defined programs that are possible with the more complex type of firmware systems.
AGAINST: Now let's look at the argument for retaining category 2 software and ignoring the GAMP 5 advice. Looking around the laboratory you'll see many common laboratory instruments (such as balances, pH meters, and dispenser dilutors) that still use a firmware chip to operate the instrument. Furthermore, the firmware cannot be changed by the user and the only way for a chip update is via the service engineer. Under these circumstances I believe that there is no need for excessive validation, and these instruments can still be qualified.
However, we now run into a slight problem. The issue is that the GAMP 5 approach is now in direct conflict with USP <1058> for analytical instrument qualification (AIQ) (3), which we discussed in detail in the last "Focus on Quality" column (4).
GAMP 5 says validate and USP <1058> says qualify. Oh dear, what shall we do?
This is a common problem when different professional groups develop guidance; the individual participants sit in a silo and fail to consider anything outside their own boundaries. Hence, when each new guidance is unveiled with a fanfare and the chink of a cash register, it is not until later that we find that due to professional myopia there are conflicts and problems. Ironically, GAMP 4 was congruent with USP <1058> and these instruments would be classified as category 2 software qualified to demonstrate their intended purpose, a simpler and quicker process than validation. Under this approach the instrument's software is implicitly validated as part of the qualification to demonstrate intended use — job done. Why do more?
POTENTIAL SOLUTION: So how do we resolve this situation? Which takes precedence — a USP general chapter or the GAMP guide? I don't think we need to phone a friend or go 50:50 to answer this question do we? So in the laboratory you'll want to ignore GAMP 5 and retain category 2 software for laboratory instruments and be consistent with USP <1058>. This approach will also make your life easier in these instances.
WARNING! Implement this approach with some caution and the application of intelligence. The reason for the warning is that most of the laboratory instruments in this section are indeed category 2 software, but some instruments can have additional functionality that needs further control (for example, theycan have programmable firmware for user-defined procedures that automate tasks over and above the basic functionality of the instrument). The ability of the users to change the function of the instrument has many advantages but also comes with a significant downside: errors and incorrect procedures. Therefore each laboratory needs to take this into consideration when deciding the software category and hence how much qualification or validation work is undertaken.
Therefore you'll need to adopt a two-stage process if you develop user-defined procedures.
As you move away from simple firmware that is implicitly tested as you qualify the instrument, you will need to take this two-stage process approach, although control of user-defined procedures can be controlled by an SOP rather than writing a validation plan every time you develop one of these user-defined procedures. You will also need to control the versions of each of these procedures and introduce change control to ensure that changes to each one are managed effectively.
The other issue that you may encounter with some instruments in this category is the incorporation of calculations in the operation of an instrument. A balance is the classic example here if it is being used for content uniformity testing. Any calculations used as part of YOUR use of the instrument should be checked as part of the qualification to ensure compliance with the requirement of 21 CFR 211.68(b) (. . . Input to and output from the computer or related system of formulas or other records or data shall be checked for accuracy. The degree and frequency of input–output verification shall be based on the complexity and reliability of the computer or related system . . .) (6). However, the calculation checks and testing should be integrated into the overall instrument qualification and not become a separate calculation validation in a multilayered validation approach.
Software Silos or Software Continuum?
The next three software categories we will discuss are intended as a continuum rather than discrete silos, so some interpretation may be necessary as to which category a system falls into. This will need to be documented in your system risk assessments or validation plans, so "think things through" is the take-home message.
Category 3 Software: What's in a Name?
Category 3 in GAMP 5 has been renamed from Standard Software to Nonconfigured Product to sharpen the difference between this and category 4 software. Now this means that software that is used as installed falls into category 3 and may (note the careful use of the word may) also include software that is configurable (category 4) but is used either unconfigured or with the standard defaults provided by the software supplier.
Despite the name, category 3 software is also configured, but for the environment (run-time configuration). It is this fact that distinguishes category 3 from category 4 software. What is run-time configuration?
As well as a simpler life cycle, which we will discuss in a later column, the software vendor or supplier's work in software development can be used to save validation effort within the laboratory, as functional and design specifications are not expected from the user. Note that this life cycle does not absolve the user from defining his requirements and also demonstrating intended, use but the testing is only a single phase. This change helps practitioners in the field to interpret software better: you could have the same software in category 3 or 4, depending whether the default settings are used or if the application is configured respectively. To illustrate this point, I have been involved in validating the same software application that was category 3 software when operating a time-of-flight mass spectrometry (TOF-MS) system (here the application was used as installed) but category 4 when used for bioanalysis with quantitative analysis using a triple-quadrupole MS instrument (here electronic signatures were configured for use in the application).
Category 4: Configured Products Refined
The name of this category has changed to help refine and redefine the software categories; the guide has moved from defining software as "package" in GAMP version 4 to "product" in GAMP version 5. I believe this is to emphasise the commercial nature of category 3 and 4 software, which constitutes the bulk of the software used in laboratories today.
The major difference between category 3 and 4 software, as mentioned earlier, is the ability to modify the function of the software to match a business process. The user has the means and knowledge to change the functionality of the device in a way that changes the results outputted by the device. As a direct consequence, this triggers increased validation effort. There are many ways to achieve this but the essence is to take standard software modules that provide the basic functionality to automate a process and change it by configuration tools. These tools are provided by the vendor of the product, hence configuration rather than using an external language to write custom code that is attached to the product. However, these tools can vary in their nature from simple configuration buttons that turn a feature on or off to graphical drag-and-drop to a modular "configuration" language that typically writes large blocks of software; hence, custom code, which raises the debate of configuration versus customization.
Understanding the difference between configuration and customization is the key to managing software and validation risk. However, many in the laboratory are seduced by suppliers marketing literature that talks of configuration when in reality it is customization, as we'll discuss now. The main point I would like to make is caveat emptor — buyer beware . . . . The user is responsible under the regulations, and if you are seduced by the marketing literature, it's your problem.
Category 4 and 5 Software: Configure versus Customize — Where Is the Line?
Configuration and customization of software are terms that are poorly defined in the validation world and frequently used interchangeably, especially in a vendor's marketing literature. It is important to understand the difference between these two terms as they mean entirely different things and consequently can have a dramatic impact on the amount of validation work that you could undertake.
The problem is that even the FDA and GAMP have not been able to define these terms, as configuration and custom are missing from the glossaries from the FDA (5) and GAMP 5 (1). However, an issue arises when a "configuration" or scripting language is provided by a software vendor to enable the modification of a program's function to fit the business process. Regardless of terminology, this language writes custom code within the application. The problem is a vendor's marketing department has typically cottoned on to the idea that custom code is a bad idea and decided to call this "configuration" instead, or worst of all, COTS software — without defining what the latter term means.
Note what GAMP 5 says on this point in Appendix M4 (1): custom software components . . . , developed with an internal scripting language, written or modified to satisfy specific user business requirements, should be treated as Category 5. In plain English this means that the so-called configuration is really customization. Therefore, any definition of customization needs to include the problem of the internal language customizations. So, here is my attempt at defining these two terms.
The two definitions discussed are very important, as you will have to determine if the software you have is customized or configured. Getting it wrong can result in generating the wrong data or receiving a noncompliance during an inspection, as we will discuss in the next section.
Category 5: Custom Macros, Modules, and Applications
Not much has changed in GAMP 5 from this apart from the name change from custom software to custom application. This is the highest risk software, as it may be unique and may not be subject to the same rigors of specification and testing as commercial products because it could be undertaken in-house or outsourced to a commercial software company.
Some further background reading on Wikipedia
However, does the name of this category really reflect the situation? The problem with this category name is the use of the word application. This implies a single application but this does not reflect the whole reality. As GAMP 5 notes (1), software is a continuum and therefore you can have a configurable product that has custom software or custom modules written to aid functionality of the installed application. A LIMS is a case in point here; there is a language provided by the LIMS vendor to change the software function — this is category 5 software as discussed in the previous section and the output is custom code. Note that the LIMS marketing people will have you think that it is configuration. No, it is custom software as it uses an internal scripting language.
This software category implicitly includes macros written say within a spectrometer application to produce a shortcut for processing or manipulating data; these macros are custom software and not an application per se. Each macro needs to be validated and controlled to ensure that it does what it is supposed to do. Also, we have in this software category the user-defined programs written on category 2 instruments; again each will be validated.
Therefore each application, module, user-defined program, or macro needs to be specified, version controlled, built, and tested (including integration testing with the commercial application, as applicable) as a minimum to ensure the quality of the software. Furthermore, the custom code modules will need to have their source code managed and backed up to prevent overwriting or loss, respectively.
Users and the Software Screwup Factor
Let us return to the discussion on software classification from a slightly different perspective. At this point I would like to introduce you to a rather vile and obnoxious four-letter word. The word is user. We now enter the realm of the user (defined as either someone in the laboratory or IT — I'm not fussy here) who has a multitude of possibilities to screw up software and also the analytical results generated by it. So we'll take another look at the GAMP software categories from the perspective of the ability of the user to change the operation of software and hence influence, manipulate, or screw up the results.
Let's summarize the impact of users on software. The further down the software classification you go, the software becomes more specific in function coupled with the ability of the software to directly influence the data and calculations and hence the final results. Or in other words, the ability of the users to screw up the software and the data increases as you get further down the software categories. Therefore, controlling and regulating the ability of users to configure or customize software is the key here. From a personal perspective, avoid writing any software unless there is a good business case for doing so. My rationale is that your function in the laboratory is analytical science, not software development.
So if you have been following the discussion so far we need to revise the GAMP 5 software categories to take account of the world in the laboratory. Presented in Table II is my modified classification of revised software categories, which is a combination of GAMP 4 with GAMP 5 classifications for the laboratory that is intended to be pragmatic and usable while managing risk effectively.
Table II: Modified software classification
First, we return to five categories of software but use or modify, in one case, the GAMP 5 titles for them. The five categories are shown in Table II together with the types of software examples found in each one.
Second, let us review each of the categories to see if and how they have been modified:
We have reviewed the GAMP 5 software categories and highlighted where there are strengths in the changes but also noted the problems with some of the categories from the laboratory perspective. I have also suggested an alternative classification that would benefit the laboratory and reinstate category 2 software. As with all my writing, the ideas and suggestions in this column are to get you to think, adopt, adapt, or reject as you see fit.
I would like to thank Lajos Hadju for his very helpful review and constructive comments during the preparation of this column. It was his input on the software classification and the impact of the users that has improved the content of this column.
R.D. McDowall is principal of McDowall Consulting and director of R.D. McDowall Limited, and "Questions of Quality" column editor for LCGC Europe, Spectroscopy's sister magazine. Address correspondence to him at 73 Murray Avenue, Bromley, Kent, BR1 3DJ, UK.
(1) Good Automated Manufacturing Practice Guidelines version 5, International Society for Pharmaceutical Engineering, Tampa, FL, 2008.
(2) Good Automated Manufacturing Practice Guidelines version 4, International Society for Pharmaceutical Engineering, Tampa, FL, 2001.
(3) United States Pharmacopoeia Inc, <1058> Analytical Instrument Qualification.
(4) R.D. McDowall, Spectroscopy 24(4), 20–27 (2009).
(5) Food and Drug Administration, Glossary of Computerized System and Software Development Terminology, 1995.
Related Content:Column: Focus on Quality