| Copyright © 2005, Eli Lilly and Company and the National Institutes of Health Chemical Genomics Center. All Rights Reserved. For more information, please review thePrivacy Policy andSite Usage and Agreement. Table of Contents - DETERMINATION OF EC50/IC50
- PRODUCTION MONITORING
DETERMINATION OF EC50/IC50 Models and Curve Fitting Guidelines For competition binding assays and functional antagonist assays the most common summary measure of the dose-response curve is the IC50, the concentration of substance that provides 50% inhibition. For agonist/stimulator assays the most common summary measure is the EC50, the concentration giving 50% of that compounds maximal response. Substantial variation in the methodology used to derive these values exists, and this variation has been shown to substantially impact overall assay variability. This section discusses important issues to consider and provides some guidelines on how to proceed. They are a based on the results data standards (Section XI). Consult that section for the specifics of each assay type. Consult a statistician to see if these guidelines are appropriate for your assay, and if other outcomes such as AUC or a threshold dose should be used. Before fitting a dose-response curve to obtain the EC/IC50, each well should be converted to either percent activity or percent inhibition with respect to positive and negative controls (note: for simplicity all text below is stated for determining IC50s; determining EC50s is identical). Then all replicate wells from a given run (including multiple plates per run) for a given concentration should be averaged either by taking the mean, or preferably, taking the median. Outliers less influence the latter when there are 3 or more replicates. Thus only one point per concentration per run is used to fit the dose-response equation to the data. This is because replicate wells on either the same or different plates are often correlated with each other and, thus, do not provide true replication of the experiment. The four parameter logistic model (4PL), also called the Hill-Slope model, is the most common equation fit to in vitro dose-response data. One form of the equation is  wherey is the percent activity andx is the corresponding concentration. The fitted IC50 parameter is therelative IC50, and is defined as the concentration giving a response half way between the fitted top and bottom of the curve. Some software, such as Activity/Base, also provides theabsolute IC50, which is defined as the concentration giving exactly a 50% response. The relative IC50 is recommended for most assays. The figures on the following page graphically demonstrate the principle of relative IC50 and relative EC50 for an inhibition and stimulation assay, respectively. A diagram for relative efficacy is also shown. 
 You should also report thefitting error, which is usually called the standard error by most software packages (we use the term fitting error to differentiate it from the standard error of the mean [SEM] derived from multiple determinations of a compound). The 4PL model is the best model for dose-response data, but there are cases where it should not be used. In some cases, due to the potency of the compound falling outside the dosing range, the data may not fully describe the bottom or top asymptote of the curve. In those cases, respectively, the bottom (3PLFB) or top (3PLFT) can be fixed to improve the curve fit. If you observe a substantial reduction in the %Fitting Error, and a better dose-response plot of the fitted curve with respect to the actual data then you should switch to either the 3PLFB or 3PLFT model as appropriate. Examples All examples below are from receptor binding data fitting %Activity versus concentration (expressed by Activity/Base as log-concentration in the plots). For this type of assay, the top, bottom and slope parameters should in theory by 100, 0 and 1 respectively.
Example 1 is a dose-response best fit by the 4PL model. Both asymptotes are defined by the data, and the fitting error is approximately the same with all 3 models. Note that even though the fitting error is smallest with the top fixed (8.63% versus 9.51%), the reduction is not small enough to warrant the fixed top model, nor is there any material change in the IC50. The fixed bottom model is clearly inappropriate as the data clearly defines a bottom >0.    | | Bottom | Top | Rel IC50 | Slope |
|---|
| 4-Param | 7.38 | 103.65 | 0.061 | -1.27 | | % Fit Err | 22.08 | 2.36 | 9.51 | -10.74 | | Bottom=0 | 0.00 | 107.10 | 0.069 | -0.96 | | % Fit Err | | 4.95 | 19.94 | -15.05 | | Top=100 | 7.79 | 100.00 | 0.066 | -1.39 | | % Fit Err | 22.07 | | 8.63 | -10.21 |
Example 1 Curve fit Results for a dose-response best fit by a 4PL model The fitting error is expressed here as a percentage of the fitted parameter value. For example, if the IC50 is 0.061 and its fitting error is 0.058, then the %Fit Error is 9.51%. Example 2 is best fit by the fixed top (3PLFT) model. The data does not define a top asymptote, and the fitted top (128.32) and slope (-0.58) from the 4PL model are inappropriate for this (binding) data. By fixing the top at 100% the fitting error is reduced from 57.54 to 21.55%, and the IC50 increases by more than two-fold. Thus the 3PLFT model should be selected over the 4PL.    | | Bottom | Top | Rel IC50 | Slope |
|---|
| 4-Param | 2.01 | 128.32 | 0.015 | -0.58 | | % Fit Err | 202.85 | 14.82 | 57.54 | -22.56 | | Bottom=0 | 0.00 | 134.33 | 0.014 | -0.53 | | % Fit Err | | 12.42 | 57.21 | -12.91 | | Top=100 | 6.21 | 100.00 | 0.034 | -0.87 | | % Fit Err | 57.08 | | 21.55 | -16.09 |
Example 2 Curve fit Results for a dose-response best fit a by a 3PLFT model Example 3 is best fit by a fixed bottom (3PLFB) model. Note that the data does not define the bottom asymptote, and the fitted bottom (41.54) and fitted slope (-1.83) from the 4PL are inappropriate for binding data. The fixed bottom model reduces the fitting error from 80.19% to 20.85%, while the IC50 increases by more than two-fold. The fitted IC50 (20.88nM) is inside the dose-range (0.001-25nM), and so it is appropriate to report this value. Note in this case Activity Base was unable to fit a fixed top model.    | | Bottom | Top | Rel IC50 | Slope |
|---|
| 4-Param | 41.54 | 106.79 | 10.17 | -1.83 | | % Fit Err | 81.91 | 2.90 | 80.19 | -95.59 | | Bottom=0 | 0.00 | 106.94 | 22.88 | -1.25 | | % Fit Err | | 2.77 | 20.85 | -30.49 | | Top=100 | | | | | | % Fit Err | | | | |
Example 3 Curve fit Results for a dose-response best fit by a 3PLFB model Example 4 illustrates the definition and effect of outliers (left panel). Outliers are single, vertically isolated points that are clearly inappropriate. The point is obviously erroneous. The effect of the outlier in this case is to bias the estimate of the bottom upwards, pulling it away from the other points of the data. In general, outliers can bias either the top, bottom or slope parameter depending upon where they occur in the dose-response. It is appropriate to remove the outlier (right panel) and refit the points. Fixing top or bottom did not materially improve the curve fit (not shown).   | All Data | Bottom | Top | Rel IC50 | Slope |
|---|
| 4-Param | 32.62 | 97.21 | 0.056 | -1.31 | | % Fit Err | 40.09 | 20.73 | 114.10 | -132.18 | | Outlier Rem | Bottom | Top | Rel IC50 | Slope |
|---|
| 4-Param | 2.25 | 104.78 | 0.130 | -0.63 | | % Fit Err | 611.06 | 11.29 | 53.38 | -40.98 |
Example 4 curve fit results for a dose-response containing an outlier Example 5 illustrates the effect of high assay variation. No single point stands out as obviously erroneous, and therefore it would be inappropriate to remove any points from the curve fit. Fixing top or bottom does not materially improve the curve fit, and so the 4PL model should be used. Note that the estimates themselves are not implausible, but the fitting error is 33.83%, which is caused by the relatively high assay variation.    | | Bottom | Top | Rel IC50 | Slope |
|---|
| 4-Param | 8.12 | 91.76 | 0.117 | -1.86 | | % Fit Err | 84.04 | 8.30 | 33.83 | -58.17 | | Bottom=0 | 0.00 | 92.98 | 0.130 | -1.51 | | % Fit Err | | 8.94 | 34.55 | -44.50 | | Top=100 | 7.62 | 100.00 | 0.093 | -1.40 | | % Fit Err | 92.45 | | 33.77 | -40.93 |
Example 5 Curve fit Results for a dose-response with high assay variability, but no outliers Notes
- This equation can be fit to the data using Activity/Base, JMP, Graphpad/Prism or Sigma/Plot. Note that the form of the equation varies from one software package to the next. Some, such as Graphpad/Prism, fit Log-IC50 instead of IC50, and the equation looks quite different, but the results are the same as that shown above.
- The termsabsolute andrelative IC50 are not universal. Both are usually just called the IC50, and its left unstated which value is actually used.
- If the software tool you are using reports Log-IC50 then you must convert both the estimate and the % fitting error (%FE) according to the formulas
 - There should be at least one point on both sides of the reported IC50, i.e. the reported IC50 should lie inside the dose-range used in the assay. The intent of this rule is to make the IC50 estimate an interpolation of generated data and not an extrapolation of generated data. Cases not satisfying this rule should not have an IC50 reported or reported with a comment that indicates the value is extrapolated. If a value is reported, it should be <X or >Y, as appropriate, where X is the lowest concentration and Y is the largest concentration included in the analysis.
- The fitting error of the IC50 should be no more than 40% of the IC50. Estimates not satisfying this rule should be flagged in the database. A fitting error of 40% of the IC50 corresponds to an MSR of 3-fold.
- It is a good idea to remove obvious outliers and then refit the curve without the outliers. Note that if it isnt obvious, it isnt an outlier. See examples 4 and 5 above to distinguish high variability from outliers.
- For competition assays, such as radioligand binding assays and competitive inhibition assays, the fitted slope should be within 2 (slope) fitting errors of the value 1, and slope estimates outside this range indicate assay problems that need to be investigated.
PRODUCTION MONITORING Production assays can be monitored in two basic ways: running control (reference) compounds and retrospective studies of compounds that have repeat evaluations that accumulate as part of the normal SAR process. Of the two methods, running control compounds is definitely better as it allows problems to be identified prospectively and corrected, whereas retrospective studies are limited to verification of past activity, be it acceptable or unacceptable. However, retrospective studies can be useful supplements, especially when conducted prior to important milestones where demonstration of valid biological assays is a requirement. Below are comments on the setup/selection of controls and the analysis of retrospective studies, and the use of bridging studies to verify that changes to assay protocols have no effect on the assay results.Control Compounds Key assays in a project and assays where problems are suspected should have two control compounds, a primary and a secondary (this is referred to asClose Monitoring). All other assays should have at least a primary control (Regular Monitoring). Both compounds need to be run once per run, unless plate variability is suspected. In that case the primary control compound needs to be run once per plate. The purpose of the primary control is to ensure that there isnt any assay drift, i.e. that the same compound has a stable Ki/Kb/EC50 over time, and that the assay reproducibility (MSR) is stable over time. The purpose of the secondary control is to examine the stability of results over a dose-range. If problems do develop, then it is important to examine whether the entire dose-range is equally affected (a small problem) or whether the dose-range is differentially affected (a big problem). Also, two controls permit direct calculation of both the within-run and overall MSRs, and a check that the MSR is consistent over a range of potencies.
The activity of the primary control should be at or near the most potent compound available, and ideally should be the Lead compound. There should also be sufficient stock of a single lot of the compound so that it can be run on a continuous basis for some period of time. Since the control compound is supposed to be representative of the test compounds, it should receive the same sample handling as all the test compounds, and not be specifically prepared and added to the assay outside of normal test compound procedures. For Close Monitoring, the secondary control should be >100 fold less potent then the primary control. Otherwise it has the same requirements as the primary control. As the SAR develops the potency traditionally improves. So when the best compounds are more than 100-fold more potent than the primary control then select a new primary control. If the assay has a secondary control then the old primary control becomes the new secondary control, and the existing secondary control is dropped. If there is no secondary control then it is suggested to run both primary controls over the first 6 runs of the new primary control. A scatter plot for control compound log-Ki/Kb/EC50 versus run date should be updated after every run and checked for problems. For assays with two control compounds the difference in log-Ki/Kb/EC50 versus run date should be plotted, and for agonist and non-competitive antagonist assays the efficacy versus run date should also be plotted. Outlier runs and trends either up or down (assay drift) should be checked visually, and problems investigated and corrected as they occur. Outlier runs should be repeated. After 6 runs compute the overall MSR of the assay based on the control compounds according to formula, , wheres is the standard deviation of the log-Ki/Kb/EC50 values. This MSR is the total oroverall MSR (whereas the one computed in a test-retest study encompasses only the within-run variability), and should be less than or equal to 7.5. This standard comes from practical experience obtained thus far with assays in the company, and not theoretical statistical considerations. Note that this is a minimum standard that all assays should meet, and in practice chemistry requirements may indicate a smaller MSR (as low as 2-3) is required for some or all assays. The Project/Program Team should discuss this issue with a statistician to set appropriate MSRs for their assays. A spreadsheet is availablehere for computation of single control per run and two controls per run protocols. See a statistician for other cases. After each run, a running MSR plot should be maintained (i.e. computed from the last 6 runs) and checked to ensure the continued good reproducibility of the assay. Examples Example 1 illustrates results for an assay with a single control. The left panel shows the potency versus run date scatter plot, the right panel the moving MSR chart. The MSR points are based on the last 6 runs of the assay, i.e. the first point is computed using runs 1-6, the second point uses runs 2-7, etc. The Mean Summary section indicates the highest/lowest/last IC50s in the period were 22.63, 4.42 and 11.25 uM respectively (chart units are in nM). The overall average was 10.17 uM. The potency has no apparent temporal trends, and no unusual observations. The right panel shows the trends in MSR over time, which appears to increase until mid Feb-2002, and then decrease. However, the magnitude of the increase trends is quite small and well within the variation of an estimate based on a sample of size 6. The highest/lowest/latest MSRs are 6.8, 2.7 and 2.7 respectively. The overall MSR is 4.4, which is not the average of the 6-run MSRs but instead is a single estimate derived using the entire sample (18 data points in this case). This is a stable assay with moderate assay variation (3 < MSR < 5).
  | Mean Summary | | MSR Summary |
|---|
| High | 22630.00 | High | 6.8 | | Low | 4420.00 | Low | 2.7 | | Overall | 10172.76 | Overall | 4.4 | | Current | 11250.00 | Current | 2.7 |
Example 1. Potency, MSR Chart, and Summary Statistics for an Assay with One Control Example 2 illustrates an assay with two controls. In the left panel the red and blue lines represent the two compounds, and are positioned using the left axis. The green line is the potency ratio between the two compounds and is positioned using the right axis. The right panel shows the moving MSR values both within run and overall. The Overall-Overall MSR is the value to be reported. The within-run MSRs are only for comparison backwards to the test-retest study results, and for times when compounds are compared within the same run of an assay. As with example 1, there are no apparent temporal problems, i.e. this is a stable assay with an overall MSR of 2.0. This assay is less variable than the assay in example 1.   | Mean Summary | | MSR Summary |
|---|
| | Ref 1 | Ref 2 | Ratio | | WR | Overall | | High | 505.26 | 10.84 | 0.02 | High | 1.9 | 2.1 | | Low | 324.19 | 3.42 | 0.01 | Low | 1.6 | 1.7 | | Overall | 409.51 | 5.53 | 0.01 | Overall | 1.8 | 2.0 | | Current | 462.99 | 7.21 | 0.02 | Current | 1.9 | 2.0 |
Example 2. Potency, MSR Chart, and Summary Statistics for an Assay with Two Controls Examples 3 and 4 illustrate problems with a shift in compound potency. Example 3 illustrates a steady degradation in potency over time, whereas Example 4 illustrates a more sudden shift in potency at a particular point in time. In Example 3 the assay variability appears to be shrinking, while in Example 4 it appears to be stationary. Repetitive freeze-thaw cycles of a compound may cause a slow degradation in potency whereas a change in lot of a key assay ingredient may result in a sudden potency shift. In both cases it is important to identify the cause and correct it as soon as possible.   Example 3. Potency and MSR Chart Illustrating Assay Drift   Example 4. Potency and MSR Chart Illustrating Sudden Change In Potency Example 5 illustrates an assay with stable potency, but in June the assay variability increased. The moving MSR was stable around 3, but after June increased to over 10, and remained there. This also is most likely caused by a change in the assay process around that time. Again it is important to identify and correct the cause as soon as possible. Note however that a single outlier will cause the MSR chart to increase for the next 6 runs, and so it usually takes more time to correctly distinguish a change in assay variability from a single outlier result.   Example 5. Potency and MSR chart for change in assay variability Retrospective Studies During the course of project/program development numerous compounds are repeatedly evaluated and stored in archival databases such as ICARIS. This data can be mined to examine the reproducibility of assay results. This work should always be done by a statistician as the repeated compounds are not a random selection of all compounds, and may be biased with respect to time of evaluation, potency, structure and assayability (the latter term is meant to reflect conditions such as solubility, quenching, stickiness to plastic and other practical problems). In spite of these potential problems retrospective studies can be a very useful exercise, particularly in establishing the acceptability of older assays that have never been formally evaluated for reproducibility. In addition, the MSR can be examined over various subsets such as potency range, structure and run date to check that the control compound MSRs are representative of the test compounds with respect to potency range, structure and run date.
Bridging Studies If a key aspect of an assay changes, such as an equipment change or lot of a reagent, then a test-retest study should be conducted to verify equivalence of the two protocols. A judgment should be made on a case-by-case basis of whether the full protocol outlined in Section II.B needs to be made, or only a single run under old and new conditions (i.e. one might do just Step 4 of the procedure, or one might do both Steps 3 and 4 depending upon the severity of the protocol change). Also in cases of specific modifications such as replacing equipment for a particular step in the assay an experiment can be designed to validate that the replacement is equivalent to the original in the conduct of that step of the assay.
Dimethylsulfoxide: biological compatibility and compound storage Dimethylsulfoxide (DMSO) is a universal solvent for all compounds tested in high, medium and low throughput screens (HTS, MTS and LTS). Compounds are initially dissolved in 100% DMSO and further diluted into water and assay buffers in subsequent dilutions for screening and IC50 or Ki determinations. It is extremely important that the DMSO compatibility of biological reagents such as enzymes, receptors, protein/peptide reagents and cells be established to ensure that the screening assays are not adversely affected. In general, the final DMSO concentrations in cell-based assays are <0.2% and are <1% in biochemical assays. It is highly recommended that the tolerable DMSO concentration be determined individually for each validated assay. DMSO is also used as a cryoprotectant in the freezing of cell cultures at ATCC. The product is cell culture grade and has been tested to ensure cell viability. Each lot is also tested for the absence of bacteria, fungi, and endotoxin. When solubilized compounds are stored in DMSO, it is important to understand the stability of these compounds under various storage conditions and freeze-thaw cycles. A detailed study of these effects was published recently (1). It is believed that the degradation of DMSO solubilized compounds is mainly due to moisture absorbed from the air. This can happen during frequent freeze-thaw cycles of compounds stored frozen in DMSO, or frequent exposure to air during repeated access for biological testing (cherry-picking). Recommended storage conditions for DMSO solubilized compounds: - 96- well polypropylene plates.
- Storage temperature: 10 degree Cor room temperature.
- Inert gas atmosphere: argon flush.
- Minimal exposure to moist environments.
REFERENCES: - Cheng X., Hochlowski J, Tan H, Hepp, D, Beckner C, Kantor S, Schmitt R, Studies on Repository Compound Stability in DMSO Under Various Conditions. J Biomol Screening, 2003, 8(3), 292-304.
- Kozikowski, BA, Burt, TM, Tirey, DA, Williams LE, . Kuzmak, BR, Stanton, DT, Morand, KL, and Nelson, SL The Effect of Room-Temperature Storage on the Stability of Compounds in DMSO J Biomol Screen 2003 8: 205-209.
|