- Research article
- Open Access
Engineering proteinase K using machine learning and synthetic genes
BMC Biotechnology volume 7, Article number: 16 (2007)
Altering a protein's function by changing its sequence allows natural proteins to be converted into useful molecular tools. Current protein engineering methods are limited by a lack of high throughput physical or computational tests that can accurately predict protein activity under conditions relevant to its final application. Here we describe a new synthetic biology approach to protein engineering that avoids these limitations by combining high throughput gene synthesis with machine learning-based design algorithms.
We selected 24 amino acid substitutions to make in proteinase K from alignments of homologous sequences. We then designed and synthesized 59 specific proteinase K variants containing different combinations of the selected substitutions. The 59 variants were tested for their ability to hydrolyze a tetrapeptide substrate after the enzyme was first heated to 68°C for 5 minutes. Sequence and activity data was analyzed using machine learning algorithms. This analysis was used to design a new set of variants predicted to have increased activity over the training set, that were then synthesized and tested. By performing two cycles of machine learning analysis and variant design we obtained 20-fold improved proteinase K variants while only testing a total of 95 variant enzymes.
The number of protein variants that must be tested to obtain significant functional improvements determines the type of tests that can be performed. Protein engineers wishing to modify the property of a protein to shrink tumours or catalyze chemical reactions under industrial conditions have until now been forced to accept high throughput surrogate screens to measure protein properties that they hope will correlate with the functionalities that they intend to modify. By reducing the number of variants that must be tested to fewer than 100, machine learning algorithms make it possible to use more complex and expensive tests so that only protein properties that are directly relevant to the desired application need to be measured. Protein design algorithms that only require the testing of a small number of variants represent a significant step towards a generic, resource-optimized protein engineering process.
Protein properties that are relevant to real-world applications are often difficult to manipulate using either of the current protein engineering paradigms [1–3]: structure-based protein design [4, 5] or directed evolution [6–8]. Both methods have shortcomings and advantages that have been discussed and compared elsewhere [1–3]. Chief amongst the limitations of both methods is the requirement for high throughput computational or physical tests to evaluate protein variants for suitability to a specific application. A common problem with both approaches is that frequently there are no high throughput tests for real applications. For example, there are no high throughput tests for measuring how well a protease will remove grass stains from jeans, how quickly an antibody will shrink a tumour, or how immunogenic a potential vaccine antigen will be. As a consequence, protein engineers are frequently forced to compromise. Thus a structure-based approach in which the effects of large numbers of amino acid changes on the active site are calculated may require the protein engineer to consider only the affinity of an enzyme for its substrate and product while ignoring the effects that temperature and solvent conditions may have on the enzyme. Similarly an empirical library based approach in which large numbers of randomly produced viral antigen variants are tested for activity may allow the protein engineer to measure their binding to antibodies already known to be neutralizing, but would prohibit direct measurement of the production of such antibodies in animals exposed to the antigens.
Many non-biotechnological engineering endeavours pose similar challenges to those found in protein engineering: a large number of independent variables and cost-prohibitions against exhaustive search. Such diverse tasks as fuel formulation, clinical trial design and chemical process optimization are solved using experimental designs to combine variables in specific ways, and regression analysis techniques to dissect out the contribution of each variable to the outcome . The common goal in all these areas of optimization is to keep the total number of activity measurements small enough to allow complex functional tests that are directly relevant to the final application.
Multivariate data analysis has been used to optimize small molecules and peptides for nearly a quarter of a century [10–16]. In their paper describing chemical synthesis of a gene in 1984, Benner and colleagues suggested that systematic variation of amino acids could provide an understanding of the relationship between a protein's sequence and its function . Until recently, however, synthesis of specifically designed individual genes has been sufficiently difficult to effectively preclude the construction of designed gene sets and meaningful testing of analytical predictions. Such efforts have thus been largely confined to the synthesis of very small numbers of discrete polynucleotide  or protein variants , or to the analysis of variants produced in a library [20–22].
A synthetic biology approach to protein engineering has been enabled by recent advances in gene synthesis technology [23–26] that permit cost-effective synthesis of individually specified gene sequences instead of relying on creation of libraries of variant sequences [27, 28]. The feasibility of producing tens or hundreds of protein variants in which all amino acid changes are precisely specified allows the sequences and activities of these variants to be analyzed using multivariate regression and machine learning techniques adapted from optimization tasks found in other engineering disciplines.
We have tested this protein engineering approach by increasing the activity and heat stability of proteinase K. We selected 24 amino acid substitutions, then designed, synthesized and tested 59 genes containing combinations of these changes. We tested 8 different machine learning algorithms for their ability to identify the amino acid changes with a beneficial effect on proteinase K activity by using them to design new variants with improved combinations of substitutions. In 3 design cycles we synthesized genes encoding a total of 95 enzymes (~100 kb of synthetic genes), some of which had 20 times higher activity than the wild type protein. All 8 algorithms produced enzyme designs that were substantially improved over wild type. The results show that machine learning models of protein sequence and activity combined with efficient gene synthesis can be valuable tools in engineering proteins with improved properties.
Results and Discussion
1. Selection of proteinase K as a test system
To test machine learning-based protein engineering we chose to optimize proteinase K-catalyzed hydrolysis of the tetrapeptide N-Succinyl-Ala-Ala-Pro-Leu p-nitroanilide following a heat-treatment of the enzyme. We selected this activity because it mimics a key characteristic of practical protein optimization; target activities frequently result from a combination of protein properties, in this case expression and post-translational processing in a heterologous host, catalytic activity and thermostability.
The gene encoding proteinase K from Tritirachium album  was re-synthesized with an E. coli codon bias [30, 31] and cloned into an arabinose-inducible E. coli expression vector. The nucleotide and amino acid sequences of this initial ("wild-type") proteinase K sequence are shown in Additional file 1.
2. Engineering proteinase K: design methods
2.1 Overview of the method
The protein engineering method described here involves the following steps.
i) Selection of amino acid substitutions to incorporate into a target protein.
ii) Design of protein variants containing different combinations of those substitutions.
iii) Synthesis of genes encoding the protein variants.
iv) Expression of the protein variants.
v) Measuring the activity of the protein variants.
vi) Analysis of protein variant sequences and activities to assess the contribution of each amino acid substitution.
vii) Design of a new set of variants using the information from vi).
viii) Iteration of steps iii) to vii).
These steps are described in detail for the engineering of proteinase K in the Results sections noted in Figure 1.
2.2 Selection of amino acid substitutions
We planned on synthesizing a total of less than 100 variants containing combinations of a limited set of amino acid substitutions. To define a search space that could be effectively explored by synthesizing such a small number of variants we chose to use twenty four amino acid substitutions within the ~370 amino acid proteinase K: less than 0.5% of the total number of single amino acid changes possible.
To select the substitutions, a set of serine proteases with >30% amino acid identity to proteinase K were identified by using the BLAST algorithm to search Genbank. This search produced 3 groups of homologous sequences. Group A contained the wild type and 5 close homologs (>90% amino acid identity). Group B contained 42 more distant homologs (between 30% and 90% amino acid identity). Group C contained 11 homologs (>30% amino acid identity) that were either reported in the literature to be thermostable or were >90% identical to a known thermostable sequence. Genbank accession numbers are provided in Additional file 1.
The homologs were aligned using clustalW, to identify the amino acids in each homolog that corresponded with the amino acid found in wild type proteinase K at each position. To increase the probability that at least some of the substitutions would increase activity, we selected 24 substitutions based on several different criteria : (i) the substitution was reported in the literature to increase the stability of the serine protease subtilisin (N95C, P97S, E138A, M145F, L299C, I310K) [34, 35]; (ii) the amino acid occurred within the homolog set (S107D); (iii) the amino acid occurred within >70% of homologs from thermophilic organisms and within other homologs (S123A, I132V, L180I, R237N, S273T, G293A, K332R, S337N); (iv) the amino acid was found within the homolog set and the substitution from the wild type residue is favourable in the Dayhoff substitution matrix (V167I, A199S, V267I) ; (v) principal component analysis of amino acids responsible for clustering of homologs from thermophilic organisms (K208H, A236V) ; (vi) literature reports that the P5S substitution has a stabilizing effect in subtilisin [34, 35] AND appearance of a P to S substitution in the closest proteinase K homolog (P265S, P355S); (vii) the change occurred in a close homolog that is also thermostable (Y151A) and (viii) a random mutation identified during synthesis of the wild type (Y194S). This information is summarized in Table 1
We emphasize that the method used for choosing amino acid substitutions is independent of the subsequent machine learning analysis. Substitutions could be selected by any of the many available methods including analysis of protein structures  or comparison of homologous sequences . A more detailed review of combined methods for substitution selection has been published elsewhere.
2.3 Design of variant set 1
In order to test the machine learning algorithms, we needed to obtain a set of variants with corresponding activity measurements. For the most accurate analysis each substitution should be approximately equally represented, and should occur with as many different substitutions (ie in different sequence contexts) as possible. We encountered two somewhat related obstacles to creating such a variant set. Firstly, all of the substitutions were previously untested, so we did not know how many would completely inactivate the enzyme. Secondly we did not know how tolerant proteinase K would be to changes, that is how many amino acids we could change in a single variant and retain activity. The initial set of 59 variants was therefore designed in several stages as we obtained information about these parameters. The sequences of all variants synthesized are shown in Additional file 2.
i) First a set of 24 variants was designed with combinations of substitutions selected randomly but with the constraint that each of the 24 substitutions occurred 6 times and each variant contained 6 substitutions (1–2 to 1–25, design method B in Additional file 2). Of these 24 variants only one (1–13) was active after heat-treatment. To determine whether the low survival rate was because the substitutions destroyed all proteinase K activity, or because the substitutions primarily affected the heat sensitivity, we also measured the activity of all 24 variants without heating. Under these less stringent conditions we found three additional variants that were active (1–2, 1–6 and 1–12). Eighteen of the 24 substitutions were present in 1 or more of these 4 variants with detectable proteinase K activity and thus did not completely inactivate the enzyme.
ii) To see which of the remaining 6 substitutions destroyed proteinase K activity, we synthesized variants containing the substitutions that had not occurred within an active variant (1–26 to 1–33, design method C in Additional file 2). Variants containing N95C, P97S, E138A, A236V and L299C were completely inactive, so these substitutions were eliminated from further designs.
iii) To obtain a larger number of active variants for modeling using machine learning we designed 10 variants by arbitrarily combining 3 substitutions that had appeared previously in active variants (1–34 to 1–43, design method D in Additional file 2) and 6 variants by arbitrarily combining 5 substitutions that had appeared previously in active variants (1–44 to 1–49, design method E in Additional file 2).
iv) Finally we performed a manual analysis of the activity data from the first 49 variants, combining substitutions that occurred frequently within active variants (1–50 to 1–59, design method F in Additional file 2).
2.4 Testing variant set 1
To associate protein sequences with functions, we tested the ability of the proteinase K variants to hydrolyze N-Succinyl-Ala-Ala-Pro-Leu p-nitroanilide. Proteinase K variants were expressed in E. coli and purified over a Ni-NTA column. The purified proteins were heated to 68°C for 5 minutes, cooled and diluted into reaction buffer containing substrate. The activities measured are shown in Figure 2 and in Additional file 2, which also show the activities of variants designed in two subsequent design cycles.
Only 19 of the 59 enzymes in variant set 1 had detectable activities after heat treatment [see Additional file 2]. As described in section 2.5, we wished to analyze this dataset using machine learning algorithms to calculate the values of 20 parameters. Using a dataset this sparse will cause inaccuracies for the machine learning algorithms. To increase the number of datapoints without increasing the number of sequences synthesized we also measured the activities of all of the enzymes in variant set 1 without the heating step. We reasoned that this would provide additional information, differentiating combinations of substitutions that eliminated enzyme activity entirely from those which were simply unable to confer thermostability. More than half (32 of 59) of the variants in set 1 were active without a heating step (Additional file 2).
2.5 Choice of machine learning algorithms and analysis of variant set 1
We wished to learn which amino acid substitutions increased activity and which were detrimental by analyzing the sequences and activities of the proteinase K variants. The initial dataset was rather sparse: only two variants in the initial set (1–40 and 1–50) had an activity exceeding that of the wild type, by 2.8-fold and 3.3-fold respectively, while 45 possessed less than 10% of the wild type activity. Despite the generally low activities of the first set of variants, there was a range of activities that we analyzed by machine learning.
To do this we first eliminated five substitutions (N95C, P97S, E138A, A236V and L299C) because variants with any of these substitutions did not have any detectable activity (Additional file 2). We then considered the reduced set of 19 substitutions, representing each variant as a 19 dimensional bit vector xi, where xi,j is 1 if there is a substitution in the variant at position j. We used a bit vector since only one possible amino acid substitution was used at each position. A test of a protein variant was encoded as a pair (xi, yi), where xi represents the variant and yi the activity measured for this variant.
We selected 8 different machine learning algorithms to analyze the data. We used 8 different algorithms because we had no way of knowing which, if any, would be suitable for analyzing protein sequences and activities. The algorithms differ in two main ways. First in the way in which they calculate the differences between the measured activity and the predicted activity (the "loss"), for example whether they use the square of the differences between measured and predicted activities (square loss), or whether they place more weight on differences between measured and predicted activities for the more active variants (matching loss). Second, the algorithms use different regularization functions, which determine for example whether preferred solutions use many small weights (2-norm) or fewer large weights (1-norm). The algorithms used were: ridge regression (RR) ; least absolute shrinkage and selection operator (Lasso) ); partial least square regression (PLSR) ; support vector machine regression (SVMR) ; linear programming support vector machine regression (LPSVMR) ; linear programming boosting regression (LPBoostR) ; matching loss regression (MR) [47, 48]; one-norm regularization matching-loss regression (ORMR) [47, 48]. See Additional file 1 for detailed descriptions of the algorithms.
Each algorithm was used to build linear models of the sequence and activity by calculating a 20-dimensional weight vector w, where the activity of a variant xj is estimated as i = (∑j = 1..19wjxi,j) + w20. The weight wj is associated with the j-th substitution, providing a measure of the effect of the j-th substitution on proteinase K activity. The last weight w20 is an additive shift. The machine learning algorithms were used to select values for wj that resulted in the best correlation between the activities that had been measured for each variant and the activities predicted by the weight vectors w. To do this we created 1000 subsamples of the training set (the set of all (xi, yi) pairs used for a cycle of machine learning) by leaving out 5 randomly chosen variant sequences for each such subsample. For all 8 algorithms we calculated the mean value and the standard deviation of each substitution weight wj over the 1000 subsamples of the training set.
2.6 Design of variant set 2
One objective in designing a second variant set was to see whether variants based on the results of machine learning analysis had improved activity relative to the training set. We also wished to obtain additional data so that we could perform a second round of machine learning-based variant design, should the first round prove successful. Variant set 2 was therefore designed in 3 parts.
Initially we used each machine learning algorithm to select the sequence that it predicted would have the highest activity using the heated activity data from the first set (variants 2-1 to 2–6, design method G in Additional file 2). The effect of any substitution may depend upon the other substitutions with which it occurs (see Section 3.7). The fewer times that a substitution has been seen, the less accurately its average effect is known. We reasoned that a substitution should be seen at least three times to estimate a meaningful average effect (if the effect in one context is positive, and in another context is negative, a third context will provide a "tiebreaker"). We therefore excluded substitutions that had occurred in active variants fewer than 3 times. To further reduce the chances of incorporating an apparently positive substitution that actually had a negative effect we only included those substitutions whose weights exceeded a threshold. We first normalized all our activities against the wild type, resulting in activity 1 for the wild type. We then chose the threshold as 0.04 = 1/25, where 25 is the original number of weights (24 substitutions plus a bias term). Note that the final number of substitutions somewhat smaller (19). This led to the exclusion of M145F, S123A, E132A and V267I from variants 2-1 to 2–6. In designing variant set 3, we improved our design method, using the standard deviation for each substitution weight instead of an arbitrary threshold (see Section 2.8).
We designed a further 14 variants in set 2 to more thoroughly explore the search space close to already tested variants and thus to provide sufficient data for a further cycle of machine learning. Six of these (2–7 to 2–12, design method H in Additional file 2) were designed using each machine learning algorithm in turn to select the variant with the largest predicted activity based on the mean weight for each substitution. To ensure that we designed sequences that were different from the first six, we only allowed variants between 3 and 5 amino acid changes from any tested variant of set 1 or any variant already chosen for inclusion in set 2. The lower bound of distance 3 assured that new and significantly different variants were chosen, and the upper bound of distance 5 limited the risk of encountering non-viable combinations.
The last 8 variants of set 2 (2–13 to 2–20, design method I in Additional file 2) were designed in the same way as 2–7 to 2–12, except that instead of using the activities after heating for the machine learning, we used the activities before heating. The reason for this was that only 19 of the 59 variants had detectable activities after heating, but 32 had detectable activities before heating. The unheated measurements thus provided a better dataset for machine learning and we reasoned that they would increase the diversity of designs of active proteinase K variants. This is discussed in more detail in section 3.6. The 0.04 threshold was not subtracted from the weights for designs H and I (our aim was to design a set of active but different variants) and the four substitutions that were excluded from 2-1 through 2–6 (M145F, S123A, E132A and V267I) were included in 2–7 through 2–20.
2.7 Testing and machine learning analysis of variant set 2
The first set of machine learning-based designs significantly outperformed those based on a manual "expert" analysis in set 1. Thirteen of the twenty variants in set 2 were more active than wild-type proteinase K, with 8 more active than the most active variant from set 1 (Figure 2 and Additional file 2]). Encouraged by this result we performed a second cycle of machine learning.
The sequence and activity data from the first and second set of variants was combined and analyzed as before using each of 8 different machine learning algorithms to build linear models of the sequence and activity as described in section 2.5. The mean weights and standard deviations calculated by each algorithm are shown in Table 2, and shown graphically for one algorithm (MR) in Figure 3A.
Because we now had more sequence and activity data, we also performed a rudimentary test of regression models that consider epistatic interactions between the selected substitutions. We did this by asking whether models containing epistatic interactions would result in a better fit between the observed and predicted activities of the variants in the training set.
Ideally we would like to know for each pair of positions (A, B) which of the 4 combinations of amino acids present at A and B maximize the activity. For 19 substitutions there are a total of 171 total possible pairs to consider (: A with B, A with C, A with D,...B with C, B with D, etc). Each pair consists of 4 possible states: both wild-type, both substituted and 2 possibilities in which only 1 is substituted. Thus to perform one separate test for each possible combination in every pair would require 684 variants. To test each of these combinations in at least 3 different sequence contexts could require >2,000 variants, which would be quite impractical. Since computational resources are cheap we instead used the 1000 subsamples of the training sets to "virtually" test which pairs of substitutions led to better predictions by our algorithms.
To do this we expanded one amino acid pair at a time into its four possible combinations and optimized the new weight vector. Thus, for each of the 171 possible position pairs (A, B), we built a model from each of the 1000 subsamples using one weight for each position except for the (A, B) pair for which we used 4 weights: one for each possible combination of the substitutions at position A and B. We computed the loss of the linear models on predicting activities of the 5 variants held out from each subsample (the loss quantifies the differences between the predicted and measured activities) and averaged the loss over the 1000 subsamples. If one or more pairs of substitutions improved the model (ie reduced the mean loss) we fixed the pair that produced models with lowest mean loss and then repeated the process. Each time we picked the pair of positions that produced the largest reduction in the average loss. We stopped expanding amino acid pairs when no further reduction of the mean loss occurred.
Examples of weights calculated by considering amino acid pairs are shown in Figure 3B. A comparison of Figure 3A with 3B shows that the expansion of 3 amino acid pairs into 4 combinations produced a model that also modified the weights of the single substitutions. Thus S123A goes from being less than 1 standard deviation above zero to more than 1 standard deviation above zero. One substitution (Y194S) goes from being very positive to negative (green circles in Figure 3). This substitution was only represented twice in active variants in the training set, and both of these variants had very low activities (variants 1–12 and 1–35; Additional file 2). The consequences of these weights for variant design are discussed in sections 3.2 and 3.3.
2.8 Design and testing of variant set 3
In designing variant set 3 we aimed to obtain further increases in proteinase K activity. We also wanted to test whether new variant designs were improved either by accounting for the general context dependence of a substitution, or by considering epistatic interactions. Variant set 3 was therefore designed in 3 parts to answer these 3 questions.
Our first two designs used only linear models. We selected one sequence for each algorithm by combining substitutions whose weights were calculated by that algorithm to be greater than zero (variants 3-1 to 3–5, design method J in Additional file 2). We selected a second sequence for each algorithm by combining substitutions whose weights were calculated by that algorithm to be at least 1 standard deviation greater than zero (variants 3–6 to 3–9, design method K in Additional file 2). Values for the mean and standard deviations for substitution weights calculated by each method are shown in Table 2.
The third design used models that considered amino acid pairs. We selected one sequence for each algorithm by combining substitutions or substitution pairs whose weights were calculated by that algorithm to be at least 1 standard deviation greater than zero (variants 3–10 to 3–16, design method L in Additional file 2). When more than one pair had a positive substitution weight, we selected the pair with the highest value when the standard deviation was subtracted from the mean. Thus in Figure 3B we chose 337S, 355P over S337N, 355P and S337N, P355S although the weights for all three combinations were more than 1 standard deviation above zero. As for designs from the linear models, we only included a combination for a pair if that combination was present in the training set at least 3 times. Thus in Figure 3B we rejected I132V, 208K because it was present only twice, but instead chose 132IK, 208H which had a slightly higher value than 132I, 208K.
For every machine learning algorithm, the design that incorporated substitutions only when the mean substitution weights were at least 1 standard deviation above zero, outperformed the design that incorporated substitutions when the mean substitution weights were simply greater than zero. There was no clear pattern when epistatic models were used: the data is shown in Figure 4 and discussed in more depth in section 3.3.
3. Analysis of the design methods
3.1 Functional contributions of amino acid substitutions
Ten of the initial set of 24 substitutions that we selected had a beneficial effect on proteinase K activity, a success rate of 40%. The substitutions were selected from a total of more than 7,000 possible (371 positions × 19 alternatives at each position), by using alignments of homologous sequences without the use of any structural information.
Table 1 shows the effect of each of substitution next to the method used to select it. Two of the 24 substitutions selected (Y151A and G293A) were selected as positive by all 8 algorithms in the second analyses (section 2.7). These substitutions were incorporated into every variant in round 3 and are present in the most active variants. Both of these substitutions were chosen from alignments of thermostable homologs of proteinase K. In all, eight of the ten positive substitutions (S123A, I132V, Y151A, L180I, S273T, G293A, K332R and S337N) were selected based on the presence of the new amino acid in an alignment of thermostable homologs. By contrast, four of the five substitutions that were not found in any active variant were designed based on literature reports of their stabilizing effects in subtilisin (N95C, P97S, E138A, and L299C)[34, 35]. We were thus most successful in choosing beneficial substitutions by selecting changes that occur in homologous natural proteins.
The positions of all substitutions used are shown mapped onto the structure of proteinase K [see Additional files 3, 4 and 5]. We could see no obvious pattern distinguishing the locations of beneficial from detrimental substitutions, nor were we able to identify simple structural reasons for the effect of the substitutions.
For future extensions of this method, if a target activity is not achieved with the initial set of substitutions, additional substitutions can be chosen and incorporated into a new set of variants along with the best substitutions that have already been tested. Results from previous experiments can improve a second cycle of substitution selection. For example, to obtain further improvements in proteinase K activity by incorporating new substitutions, we would pick more substitutions that appear in alignments of thermostable homologs and avoid those reported to confer stability on subtilisin. More data from other systems will be required to determine whether the best method for picking substitutions varies depending on the protein target or the desired application. In either case, using a variety of methods for the initial selection, then analyzing the functional contributions of substitutions selected by different methods is likely to provide a good starting point for other protein engineering projects.
3.2 Representation of substitutions in the training data set
In our initial set of variant designs (section 2.3) we aimed to have each amino acid substitution represented more than 6 times. Because so many of our random combinations of substitutions were inactive this resulted in a training set where different substitutions were represented very unevenly.
There are two major consequences of underrepresentation of a substitution for machine learning analysis. One can be seen in Figure 3A, where the MR algorithm assigned Y194S a high weight even though both variants in the training set have very low activity. Because there are only 2 active variants encoding Y194S, the machine learning algorithms tend to assign weights to the substitution that improve the fit of other substitutions to the model, but do not really reflect the contribution of the underrepresented substitution. The more the substitution is represented the less likely this is to occur because there are more datapoints that the weight has to be consistent with. Table 2 also shows that this phenomenon is dependent on the machine learning algorithm used. Two of the three algorithms that use one-norm regularization (Lasso and LPSVMR) and use fewer larger weights to fit the data give very low scores to Y194S.
A second consideration that arises when a substitution is underrepresented is that it is difficult to assess effects of context upon the contribution of a substitution. Sometimes a substitution may be beneficial with one set of other substitutions, but deleterious with a different set. The fewer times a substitution has been tested, the less likely such interactions are to be detected.
In this study we required that a substitution occur at least 3 times for us to use it in a subsequent design. For future designs of variant sets it will be important to ensure that each substitution is adequately represented in the training data set.
3.3 Accounting for interactions between amino acid substitutions
The extent to which one substitution affects the contribution of another substitution to protein function is difficult to predict. For proteins that have evolved by the sequential accumulation of point mutations, most of those mutations must work well in many different contexts. This is because each new mutation will produce a new sequence context, so an enzyme whose amino acids were predominantly very context dependent would be largely immutable. This view is supported by a study in which all amino acid differences in 15 natural subtilisins were recombined by DNA shuffling . Almost all possible pairwise combinations of amino acid differences were found in functional subtilisin enzymes produced by this recombination, suggesting that amino acid covariation seen in the original 15 orthologs resulted from common ancestral derivation rather than functional constraints. Selecting amino acid substitutions that occur in natural homologs should therefore provide a useful bias towards variations that are tolerated in many contexts.
Different subsamples of the training set produced different values for the weights of each substitution. This difference probably arises from noise in the data as well as from possible context effects. To accommodate this variation in our designs we used the standard deviation of each weight as a measure of its variability of effect. We compared the activities of third cycle variants designed by combining all substitutions whose mean weights were positive (substitution weight example shown in Figure 3A blue and red circles, activities shown in Figure 4 variants 3-1 to 3–5, green bars), with those designed by combining only substitutions whose mean weight was more than 1 standard deviation above zero (substitution weight example shown in Figure 3A red circles only, activities shown in Figure 4 variants 3–6 to 3–9, red bars). For every machine learning algorithm, the variant that contained only substitutions whose mean weights were at least 1 standard deviation above zero was more active than the corresponding variant that included all substitutions with positive weights. The standard deviation of a substitution weight thus appears to provide a useful evaluation of the likely contribution of that substitution to protein function.
As described in sections 2.7 and 2.8 we also tested designs based on regression models that considered epistatic amino acid interactions. Activities of variants designed in this way are also shown in Figure 4 (variants 3–10 through 3–16, blue bars, substitution weight example in Figure 3B). Only the algorithms PLSR and LPBoostR produced more active designs based on modeling amino acid interactions than the corresponding designs produced when all substitutions were modeled independently. One of these, LPBoostR, found the most active of all the 95 sequences we tested (variants 3–11 and 3–12). However we note that different machine learning algorithms selected different amino acid pairs for expansion. It is therefore unclear to us whether these pairings are actually related to epistatic interactions between the amino acids themselves, or result from differences in the machine learning methods' ways of minimizing discrepancies between measured and predicted activities in a small and unevenly distributed dataset.
Understanding interactions between specific pairs of amino acid substitutions is unlikely to limit the protein engineering method described here. To test every combination of all pairs of amino acid substitutions would rapidly become prohibitively expensive: for 19 substitutions it would require more than 2000 variants (see section 2.7). However it is relatively simple to instead select substitutions that work well in many contexts and to reject those that work well in some contexts but poorly in others. This can be done by using the standard deviation of the substitution weight over many subsamples of the training set, keeping only those whose mean weights are more than one standard deviation above zero. This will also ensure that if additional substitutions are incorporated subsequently, those substitutions already accepted and fixed are likely to be generally tolerant to further change.
3.4 Comparison with other design methods
As a control to determine whether the same degree of activity improvement could be achieved by simpler means, we analyzed the activity distribution for 4 sets of variants. The first set, taken from the first 49 variants synthesized, comprised 20 variants which contained arbitrarily selected combinations of the 19 substitutions considered in the machine learning designs (1–2, 1–6, 1–12, 1–13 and 1–26 through 1–49; Figure 5, white bars). The second set comprised 10 variants that were designed by our "expert" analysis of the sequence and activity data from the first 49 variants (1–50 through 1–59; Figure 5, light shading). The third and fourth sets comprised 20 and 16 variants designed using machine learning analysis (2-1 through 3–16; Figure 5, dark shading and black fill respectively). The activities of the randomly designed variants are predominantly extremely low: 80% are less than 3% of wild type activity and just one is more active than wild type. The activities for the variants designed by manual data analysis are a little more evenly distributed, but still only one is more active than wild type. By comparison 70% of the variants designed in the first cycle of machine learning were more active than wild type and all of the variants designed in the second cycle of machine learning were at least 3-fold more active than wild type. While it is not possible for us to compare machine learning with all available protein engineering methods, this control shows that machine learning identified highly functional combinations of substitutions that could not be readily obtained either by random selection or by manual analysis.
The machine learning designs, which resulted in enzyme activity increases of up to 20-fold, differed from classical experimental designs (see for example ) because of the epistatic effect of some amino acid changes. Amino acid changes or pairs of changes that completely eliminate protein activity will mask any positive or negative contributions made by other substitutions that occur with them. Experimental designs such as Taguchi matrices [50, 51] minimize the number of experiments by combining many variables at once. A Taguchi orthogonal design for testing 24 substitutions would have produced 48 variants containing different combinations of 12 substitutions and 1 containing all 24. Our initial design incorporated only 6 changes into each of 24 variants. Although this was a less complete testing of the combinations, because 5 substitutions abolished enzyme activity, we did obtain 4 variants with detectable activity. By synthesizing an additional 8 variants we were able to identify the 19 functional amino acids in our set of substitutions. By contrast the Taguchi design would almost certainly have produced only inactive variants and thus no information.
Machine learning has also been used in the related domain of drug design to search large libraries of small molecules for compounds with maximal activity towards a biological target. The activity levels of some compounds are known and "active learning" methods  are used to select the next batch of compounds to be tested . There are a number of significant differences between these searches and those in a protein engineering setting. For example small molecules are described by feature vectors of sizes between 10 and 105 , while each protein variant in this study is described by 24 binary features. Another difference is that in drug discovery large datasets are available for testing various machine learning methods [53, 55] while no such data exists for proteins. Small molecule datasets are also generally quite large, typically with 103 to 104 compounds: Fang et al estimate that training sets of 10,000 member compounds are required to build a predictive model but in this study we tested a total of less than 100 variant proteins. In part this is possible because our protein descriptors are so much simpler and the relatedness of any pair of variants is unambiguous. This allowed us to identify improved proteins and then to focus on highly related proteins.
3.5 Multiple protein properties are modified simultaneously
The activity of proteinase K that we targeted depends on activity towards the substrate and heat-stability of the protein. We were interested in knowing whether we had modified one or both of these properties. A second motivation was that we were unable to measure the concentration or proteinase K: it autodigested so efficiently that we were unable even to visualize it on a gel. Since the half-life of the protein at 68°C should be essentially independent of the protein concentration, changes in half-life reflect changes in the protein itself and not possible influences of expression levels.
We measured the activity towards the substrate and the half-life at 68°C of 13 of the best variants. Figure 6A shows the activity of wild type proteinase K following different exposures to 68°C, Figure 6B shows one of the third cycle variants (3–9) after the same heating times. Figure 7 shows the activity without heating (white bars) and the half-life (shaded bars) for wild type proteinase K and 13 third cycle variants, as well as the substitutions in each variant. With combinations drawn only from a small set of 24 selected substitutions, a significant diversity of functional combinations provided the desired outcome, from variants in which the primary effect was increasing overall activity (3-3) to those in which both activity and half-life were improved (3–11). We expect that this pattern would continue if we attempted to deconvolute further. For example, the increase in activity without heat treatment is probably a combination of increased specific activity and increased protein expression levels, with varying contributions from each activity in each variant. Most importantly for the approach described here, several properties were altered simultaneously to improve an activity that depended on multiple properties.
3.6 Different machine learning predictions from different data sets
Different parts of variant set 2 were designed using different data sets (see section 2.6). Variants 2-1 to 2–12 were designed using the activity of the first set of protein variants after heat-treatment, while variants 2–13 to 2–20 used the activity without heat treatment. Variants 2–13 to 2–20 contained new combinations of the substitutions incorporated into 2-1 to 2–12, as well as 2 that were not included using only the heated data (S273T and V167I; Additional file 2). Five variants designed using the unheated data were more active than wild type, approximately the same proportion as those that were designed using the activities after heating (Additional file 2).
Although some variants designed using the unheated activities were active after heating, the activities without heating were in no way intended as a surrogate for the activity after heating. We used the unheated activities only to obtain a larger set of sequences, but did not measure the activities of variant sets 2 or 3 without heating. There are clearly combinations of substitutions that produce increased activity over wild type without heat treatment, but lower activity than wild type after heating (eg 1–13, 1–30, 1–37, 1–43 and 1–47). If we had performed several cycles of engineering using only the unheated activities, we would therefore expect only a subset to be active after heating.
3.7 Differences in predictions of the machine learning algorithms
The different machine learning algorithms did not converge to the same sequence design. The eight algorithms produced 4 different variant designs (3–6 through 3–9) using the same training data set. These differences in turn arose from differences in calculated mean and standard deviations for the substitution weights (shown in Table 2), which themselves resulted from differences in the way in which the algorithms model the data. The activities of the variants designed using different machine learning algorithms were very comparable (see Figure 4), and we were unable to really distinguish between them by their performances. The comparable performance of all the machine learning algorithms we used is probably due to the fact that we have too few example proteins. We expect that with more examples, clear differences between the algorithms could appear. Testing this hypothesis will require analysis of additional datasets.
It is unclear from the activity data whether there is a single optimal sequence, although there are clearly many improved sequences. For example the two most thermostable variants, 3–9 and 3–11, share 3 substitutions but differ at 5 positions (see Figure 7). The substitutions I132V and L180I appear in 3–9 but not in 3–11. Addition of either of these 2 substitutions to 3–11 leads to variants with lower thermostability than either 3–9 or 3–11 (variants 3–2 and 3–12). Thus the effect of these two substitutions appears to be influenced by other changes in the protein. This context dependence of substitutions suggests a limitation for approaches such as site saturation mutagenesis , in which all changes are considered independently and then combined based only on their behaviour in the wild-type context.
We have developed a new synthetic biology approach to protein engineering in which amino acid substitutions are selected, incorporated in different but defined combinations into a small number of variant enzymes which are individually synthesized and tested functionally. Machine learning algorithms are then used to assign values to the functional contribution of each substitution, which serves as the basis for a further set of variant designs. The process is repeated until a target activity is achieved. We have tested the approach using proteinase K as a target protein.
Substitutions that improved the activity of proteinase K were primarily identified using alignments of naturally occurring homologous proteins, structural information was not used. The exponential accumulation of natural DNA sequences [57, 58] could facilitate the use of phylogenetic information for substitution selection in many other systems, helping to remove the prerequisite of obtaining high resolution crystal structures before initiating a protein engineering project.
We tested 8 different machine learning algorithms and found them all able to produce predictive models describing the contributions of individual amino acid substitutions to the activity of proteinase K. We also found that it was unnecessary to consider all possible amino acid interactions to obtain substantial improvements in protein activity. However, it was advantageous to use the machine learning models to identify (and eliminate) substitutions whose effect appeared to vary significantly depending on the sequence context.
By designing, synthesizing and testing a total of only 95 specific proteinase K variants, of which 36 were designed using machine learning algorithms, we obtained a 20-fold increase in protein activity. Application of the strategy described here to other systems should allow proteins to be optimized using functional measurements for small numbers of protein variants. This would obviate the need for library construction and high throughput screening. Instead, variants could be directly tested in complex low-throughput assays that accurately reflect the combination of properties desired for the final application of the optimized protein.
A proteinase K-encoding gene was designed using Protein-2-DNA software  to select a codon distribution mimicking natural highly expressed E. coli proteins . The gene was assembled from chemically synthesized oligodeoxyribonucleotides (purchased from Operon) as described previously [59, 60], and cloned into pBAD/gIII (Invitrogen). Variant genes were synthesized by replacing oligonucleotides encoding the amino acids with those encoding the desired amino acid substitutions. Variants were cloned between the NcoI and SalI sites of the sequence [see Additional file 1].
Proteinase K expression and purification
Proteinase K was expressed in the E. coli periplasm and purified on Ni-NTA. Briefly, a single colony of E. coli carrying a proteinase K variant in pBAD/gIII was picked from a carbenicillin plate and grown overnight. Forty microliters of culture was then diluted into 4 ml pre-warmed LB-carbenicillin, grown for 3 to 4 hours (to an A600 of 0.2–0.3), arabinose was added to 0.2% w/v and the cells were grown overnight. All growth was performed at 30°C in LB containing 50 μg/ml carbenicillin. Cells were pelleted in a microfuge, thoroughly resuspended in 200 μl of 20% wv sucrose, 200 mM NaH2PO4, pH 7.4, 1 mM EDTA and 30 U/μl lysozyme (freshly added from a 30,000 U/μl stock: Ready-Lyse, Epicentre) and incubated at 25°C for 5 minutes. Cells were subjected to osmotic shock by addition of 200 μl ice-cold water, mixed by inversion and incubated for 5 minutes on ice to release periplasmic protein. Cells were pelleted in a microcentrifuge. The supernatant was removed, adjusted to 300 mM NaCl, 10 mM imidazole, 67 mM NaH2PO4, pH 7.4 and loaded onto an Ni-NTA column (Qiagen) pre-equilibrated with 50 mM NaH2PO4, pH 7.4, 300 mM NaCl, 10 mM imidazole. The column was washed twice with 600 μl of 50 mM NaH2PO4, pH 7.4, 300 mM NaCl, 20 mM imidazole and then eluted twice with 100 μl of 50 mM Tris-Cl pH 7.4, 300 mM NaCl, 250 mM imidazole. The two eluates for each culture were pooled.
Proteinase K activity measurements
Proteinase K Ni-NTA eluates were heat-treated in a PCR machine at 68°C for 5 minutes. Proteinase K activity was measured by addition of 10 μl Ni-NTA eluate to 90 μl reaction buffer. Final reaction conditions were 50 mM Tris-Cl pH 7.4, 180 mM NaCl, 5 mM CaCl2, 25 mM imidazole, 500 μM N-Succinyl-Ala-Ala-Pro-Leu p-nitroanilide substrate (Sigma S-8511). The reaction was incubated at 37°C, and followed by measuring absorbance at 405 nm. Activities were calculated from the initial rate of reaction and comparison with a standard curve constructed using 4-nitroanilide [61–63]. Because proteinase K self-digests we were unable to accurately determine protein concentration. Activities are therefore expressed either relative to the wild-type enzyme activity, or as pmol substrate hydrolyzed per second per ml of initial culture from which the proteinase K variant was purified (pmol/s/ml). The activities measured for each sequence in Set 1 and Set 2 are shown in Additional file 2.
Proteinase K half-life calculations
Proteinase K Ni-NTA eluates were heat-treated in a PCR machine at 68°C for 0, 2.5, 5, 7.5, 10 and 15 minutes. Proteinase K activity was measured at 37°C, and followed by measuring absorbance at 405 nm. Initial reaction rates were plotted against heating time and fitted to an exponential curve.
Machine learning algorithms
Machine learning algorithms and methods used for variant design are detailed in Additional file 1. Variant set 1 (active variants in Additional file 2) was the training data set used for the design of set 2. Multiple measurements of the same variant were treated as separate pairs. All algorithms were implemented using commercially available software (Matlab from Mathworks).
van Regenmortel MH: Are there two distinct research strategies for developing biologically active molecules: rational design and empirical selection?. J Mol Recognit. 2000, 13 (1): 1-4. 10.1002/(SICI)1099-1352(200001/02)13:1<1::AID-JMR490>3.0.CO;2-W.
Ryu DD, Nam DH: Recent progress in biomolecular engineering. Biotechnol Prog. 2000, 16: 2-16. 10.1021/bp088059d.
Tobin MB, Gustafsson C, Huisman GW: Directed evolution: the 'rational' basis for 'irrational' design. Curr Opinion on Structural Biology. 2000, 10: 421-427. 10.1016/S0959-440X(00)00109-3.
Korkegian A, Black ME, Baker D, Stoddard BL: Computational thermostabilization of an enzyme. Science. 2005, 308 (5723): 857-860. 10.1126/science.1107387.
Dwyer MA, Looger LL, Hellinga HW: Computational design of a biologically active enzyme. Science. 2004, 304 (5679): 1967-1971. 10.1126/science.1098432.
Roberts RW: Totally in vitro protein selection using mRNA-protein fusions and ribosome display. Curr Opin Chem Biol. 1999, 3 (3): 268-273. 10.1016/S1367-5931(99)80042-8.
Crameri A, Raillard SA, Bermudez E, Stemmer WPC: DNA shuffling of a family of genes from diverse species accelerates directed evolution. Nature. 1998, 391: 288-291. 10.1038/34663.
Ness JE, Kim S, Gottman A, Pak R, Krebber A, Borchert TV, Govindarajan S, Mundorff EC, Minshull J: Synthetic shuffling expands functional protein diversity by allowing amino acids to recombine independently. Nat Biotechnol. 2002, 20 (12): 1251-1255. 10.1038/nbt754.
Atkinson AC, Donev AN: Optimum Experimental Designs. Oxford Statistical Science Series. 1992, Oxford , Clarendon Press
Eriksson L, Jonsson J, Hellberg S, Lindgren F, Skagerberg B, Sjostrom M, Wold S: Peptide QSAR on substance P analogues, enkephalins and bradykinins containing L- and D-amino acids. Acta Chem Scand. 1990, 44: 50-55.
Hellberg S, Sjostrom M, Skagerberg B, Wold S: Peptide quantitative structure-activity relationships, a multivariate approach. J Med Chem. 1987, 30 (7): 1126-1135. 10.1021/jm00390a003.
Hellberg S, Sjöström M, Wold S: The prediction of bradykinin potentiating potency of pentapeptides. An example of a peptide quantitative structure-activity relationship. Acta Chem Scand B. 1986, 40: 135-140.
Mee RP, Auton TR, Morgan PJ: Design of active analogues of a 15-residue peptide using D-optimal design, QSAR and a combinatorial search algorithm. J Pept Res. 1997, 49: 89-102.
Norinder U, Rivera C, Unden A: A quantitative structure-activity relationship study of some substance P-related peptides. A multivariate approach using PLS and variable selection. J Pept Res. 1997, 49 (2): 155-162.
Sandberg M: Deceiphering sequence data, a multivariate approach. Dept Organic Chemistry. 1997, Umea , Umea University
Strom MB, Haug BE, Rekdal O, Skar ML, Stensen W, Svendsen JS: Important structural features of 15-residue lactoferricin derivatives and methods for improvement of antimicrobial activity. Biochem Cell Biol. 2002, 80 (1): 65-74. 10.1139/o01-236.
Nambiar KP, Stackhouse J, Stauffer DM, Kennedy WP, Eldredge JK, Benner SA: Total synthesis and cloning of a gene coding for the ribonuclease S protein. Science. 1984, 223 (4642): 1299-1301. 10.1126/science.6322300.
Jonsson J, Norberg T, Carlsson L, Gustafsson C, Wold S: Quantitative sequence-activity models (QSAM) - tools for sequence design. Nucleic Acids Res. 1993, 21: 733-739. 10.1093/nar/21.3.733.
Bucht G, Wikstrom P, Hjalmarsson K: Optimising the signal peptide for glycosyl phosphatidylinositol modification of human acetylcholinesterase using mutational analysis and peptide-quantitative structure-activity relationships. Biochim Biophys Acta. 1999, 1431 (2): 471-482.
Aita T, Hamamatsu N, Nomiya Y, Uchiyama H, Shibanaka Y, Husimi Y: Surveying a local fitness landscape of a protein with epistatic sites for the study of directed evolution. Biopolymers. 2002, 64 (2): 95-105. 10.1002/bip.10126.
Aita T, Iwakura M, Husimi Y: A cross-section of the fitness landscape of dihydrofolate reductase. Protein Eng. 2001, 14 (9): 633-638. 10.1093/protein/14.9.633.
Aita T, Uchiyama H, Inaoka T, Nakajima M, Kokubo T, Husimi Y: Analysis of a local fitness landscape with a model of the rough Mt. Fuji-type landscape: application to prolyl endopeptidase and thermolysin. Biopolymers. 2000, 54 (1): 64-79. 10.1002/(SICI)1097-0282(200007)54:1<64::AID-BIP70>3.0.CO;2-R.
Tian J, Gong H, Sheng N, Zhou X, Gulari E, Gao X, Church G: Accurate multiplex gene synthesis from programmable DNA microchips. Nature. 2004, 432 (7020): 1050-1054. 10.1038/nature03151.
Kodumal SJ, Patel KG, Reid R, Menzella HG, Welch M, Santi DV: Total synthesis of long DNA sequences: synthesis of a contiguous 32-kb polyketide synthase gene cluster. Proc Natl Acad Sci U S A. 2004, 101 (44): 15573-15578. 10.1073/pnas.0406911101.
Xiong AS, Yao QH, Peng RH, Li X, Fan HQ, Cheng ZM, Li Y: A simple, rapid, high-fidelity and cost-effective PCR-based two-step DNA synthesis method for long gene sequences. Nucleic Acids Res. 2004, 32 (12): e98-10.1093/nar/gnh094.
Young L, Dong Q: Two-step total gene synthesis method. Nucleic Acids Res. 2004, 32 (7): e59-10.1093/nar/gnh058.
Chen KQ, Arnold FH: Enzyme engineering for nonaqueous solvents: random mutagenesis to enhance activity of subtilisin E in polar organic media. Biotechnology (N Y). 1991, 9 (11): 1073-1077. 10.1038/nbt1191-1073.
Stemmer WP: DNA shuffling by random fragmentation and reassembly: in vitro recombination for molecular evolution. Proc Natl Acad Sci U S A. 1994, 91 (22): 10747-10751. 10.1073/pnas.91.22.10747.
Gunkel FA, Gassen HG: Proteinase K from Tritirachium album Limber. Eur J Biochem. 1989, 179: 185-194. 10.1111/j.1432-1033.1989.tb14539.x.
Gustafsson C, Govindarajan S, Minshull J: Codon bias and heterologous protein expression. Trends Biotechnol. 2004, 22 (7): 346-353. 10.1016/j.tibtech.2004.04.006.
Henaut A, Danchin A: Analysis and predictions from Escherichia coli sequences. Escherichia coli and Salmonella typhimurium cellular and molecular biology. Edited by: Neidhardt F C, Curtiss RIII, Ingraham J, Lin E, Brooks Low K, Magasanik B, Reznikoff W, Riley M, M. S, Umbarger H. 1996, Washington, D.C , ASM press, 2: 2047-2066.
Thompson JD, Higgins DG, Gibson TJ: CLUSTAL W: improving the sensitivity of progressive multiple sequence alignment through sequence weighting, position-specific gap penalties and weight matrix choice. Nucleic Acids Res. 1994, 22 (22): 4673-4680. 10.1093/nar/22.22.4673.
Ness JE, Cox AJ, Govindarajan S, Gustafsson C, Gross RA, Minshull J: Empirical biocatalyst engineering: escaping the tyranny of high throughput screening. ACS Symposium Series 900Polymer Biocatalysis and Biomaterials. Edited by: Cheng HA, Gross RA. 2005, Washington, DC , American Chemical Society, 900: 37-50.
Almog O, Gallagher DT, Ladner JE, Strausberg S, Alexander P, Bryan P, Gilliland GL: Structural basis of thermostability. Analysis of stabilizing mutations in subtilisin BPN'. J Biol Chem. 2002, 277 (30): 27553-27558. 10.1074/jbc.M111777200.
Bryan PN: Protein engineering of subtilisin. Biochim Biophys Acta. 2000, 1543 (2): 203-222.
Dayhoff MO, Eck FV: A Model of Evolutionary Change in Proteins. Atlas of Protein Sequence and Structure. 1968, 3: 33-41.
Casari G, Sander C, Valencia A: A method to predict functional residues in proteins. Nat Struct Biol. 1995, 2: 171-178. 10.1038/nsb0295-171.
Schoch GA, Attias R, Le Ret M, Werck-Reichhart D: Key substrate recognition residues in the active site of a plant cytochrome P450, CYP73A1. Homology guided site-directed mutagenesis. Eur J Biochem. 2003, 270 (18): 3684-3695. 10.1046/j.1432-1033.2003.03739.x.
Lehmann M, Kostrewa D, Wyss M, Brugger R, D'Arcy A, Pasamontes L, van Loon AP: From DNA sequence to improved functionality: using protein sequence comparisons to rapidly design a thermostable consensus phytase. Protein Eng. 2000, 13 (1): 49-57. 10.1093/protein/13.1.49.
Minshull J, Ness JE, Gustafsson C, Govindarajan S: Predicting enzyme function from protein sequence. Curr Opin Chem Biol. 2005, 9 (2): 202-209. 10.1016/j.cbpa.2005.02.003.
Hoerl AE, Kennard RW: Ridge regression: Biased estimation for nonorthogonal problems. Technometrics. 1970, 12: 55-67. 10.2307/1267351.
Tibshirani R: Regression selection and shrinkage via the lasso . J Royal Statist Soc B. 1996, 58: 267-288.
Wold H: Estimation of principal components and related models by iterative least squares. Multivariate Analysis. Edited by: Krishnaiaah PR. 1966, New York , Academic Press, 391-420.
Drucker H, Burges C, Kaufman L, Smola A, Vapnik V: Support Vector Regression Machines. Neural Information Processing Systems. Edited by: Moser M, Jordan J, Petsche T. 1997, MIT Press, 9: 155-161.
Smola AJ, Schölkopf B: A Tutorial on Support Vector Regression. Technical Report Series in Neural and Computational Learning. 1998, London , Royal Holloway College, University of London, UK,
Demiriz A, Bennett KP, Shawe-Taylor J: Linear Programming Boosting via Column Generation. Machine Learning. 2001, 46: 225-254. 10.1023/A:1012470815092.
Helmbold DP, Kivinen J, Warmuth MK: Worst-case loss bounds for single neurons. Advances in Neural Information Processing Systems. Edited by: Touretzky DS, Mozer M, Hasselmo ME. 1995, Cambridge, MA , MIT Press, 8: 309-315.
Liao J: . 2005, Santa Cruz , University of Santa Cruz, PhD:
Govindarajan S, Ness JE, Kim S, Mundorff EC, Minshull J, Gustafsson C: Systematic variation of amino Acid substitutions for stringent assessment of pairwise covariation. J Mol Biol. 2003, 328 (5): 1061-1069. 10.1016/S0022-2836(03)00357-7.
Taguchi G: Introduction to quality engineering. 1986, Dearborn, MI , Asian Productivity Organization (Distributed by American Supplier Institute Inc)
Taguchi G, Chowdhury S, Wu Y: Taguchi's Quality Rngineering Handbook. 2004, John Wiley & Sons Inc
Mitra P, Murthy CA, Pal SK: A probabilistic active support vector learning algorithm. IEEE Trans Pattern Anal Mach Intell. 2004, 26 (3): 413-418. 10.1109/TPAMI.2004.1262340.
Warmuth MK, Liao J, Ratsch G, Mathieson M, Putta S, Lemmen C: Active learning with support vector machines in the drug discovery process. J Chem Inf Comput Sci. 2003, 43 (2): 667-673. 10.1021/ci025620t.
Lam RL, Welch WJ: Comparison of methods based on diversity and similarity for molecule selection and the analysis of drug discovery data. Methods Mol Biol. 2004, 275: 301-316.
Fang J, Dong Y, Lushington GH, Ye QZ, Georg GI: Support vector machines in HTS data mining: Type I MetAPs inhibition study. J Biomol Screen. 2006, 11 (2): 138-144. 10.1177/1087057105284334.
Kretz KA, Richardson TH, Gray KA, Robertson DE, Tan X, Short JM: Gene site saturation mutagenesis: a comprehensive mutagenesis approach. Methods Enzymol. 2004, 388: 3-11.
Goldberg SM, Johnson J, Busam D, Feldblyum T, Ferriera S, Friedman R, Halpern A, Khouri H, Kravitz SA, Lauro FM, Li K, Rogers YH, Strausberg R, Sutton G, Tallon L, Thomas T, Venter E, Frazier M, Venter JC: A Sanger/pyrosequencing hybrid approach for the generation of high-quality draft assemblies of marine microbial genomes. Proc Natl Acad Sci U S A. 2006, 103 (30): 11240-11245. 10.1073/pnas.0604351103.
Venter JC, Remington K, Heidelberg JF, Halpern AL, Rusch D, Eisen JA, Wu D, Paulsen I, Nelson KE, Nelson W, Fouts DE, Levy S, Knap AH, Lomas MW, Nealson K, White O, Peterson J, Hoffman J, Parsons R, Baden-Tillson H, Pfannkoch C, Rogers YH, Smith HO: Environmental genome shotgun sequencing of the Sargasso Sea. Science. 2004, 304 (5667): 66-74. 10.1126/science.1093857.
Cello J, Paul AV, Wimmer E: Chemical synthesis of poliovirus cDNA: generation of infectious virus in the absence of natural template. Science. 2002, 297: 1016-1018. 10.1126/science.1072266.
Ciccarelli RB, Gunyuzlu P, Huang J, Scott C, Oakes FT: Construction of synthetic genes using PCR after automated DNA synthesis of their entire top and bottom strands. Nucleic Acids Res. 1991, 19 (21): 6007-6013. 10.1093/nar/19.21.6007.
Del Mar EG, Largman C, Brodrick JW, Fassett M, Geokas MC: Substrate specificity of human pancreatic elastase 2. Biochemistry. 1980, 19 (3): 468-472. 10.1021/bi00544a011.
Kasafirek E, Fric P, Slaby J, Malis F: p-Nitroanilides of 3-carboxypropionyl-peptides. Their cleavage by elastase, trypsin, and chymotrypsin. Eur J Biochem. 1976, 69 (1): 1-13. 10.1111/j.1432-1033.1976.tb10852.x.
Santos CF, Paula CA, Salgado MC, Oliveira EB: Kinetic characterization and inhibition of the rat MAB elastase-2, an angiotensin I-converting serine protease. Can J Physiol Pharmacol. 2002, 80 (1): 42-47. 10.1139/y02-004.
Swiss Protein Data Bank Viewer. [http://swissmodel.expasy.org/spdbv/]
This research was supported in part by a subcontract to DNA 2.0 from DARPA grant W911QY-04-C-0082 to Richard Gross at Polytechnic University. The authors thank Professor Gross for valuable discussions. We also thank Elaina ten Bosch, Danny Kay and Lisa Pimentel for excellent technical assistance. We are grateful to Hava Fey for support, advice and encouragement.
JL performed the machine learning analyses and new variant designs. MKW guided the machine learning analysis and new variant designs and helped to draft the manuscript. SG selected the initial amino acid substitutions, designed the first set of variants, designed synthesis strategies for all variants and participated in development of new variant designs. JEN synthesized genes and developed the protein purification and assay. RPW synthesized genes and analyzed their sequences. CG participated in experimental designs and in development of new variant designs and helped to draft the manuscript. JM synthesized genes, performed activity measurements, participated in experimental designs and in development of new variant designs and drafted the manuscript. All authors read and approved the final manuscript.
Electronic supplementary material
Additional File 2: Table 2. Contains sequence and activity information for all variants tested in this study. (DOC 489 KB)
Additional File 4: Supporting Material Figure 3. Positions of amino acid substitutions mapped onto the structure of proteinase K. (PPT 128 KB)
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
About this article
Cite this article
Liao, J., Warmuth, M.K., Govindarajan, S. et al. Engineering proteinase K using machine learning and synthetic genes. BMC Biotechnol 7, 16 (2007). https://doi.org/10.1186/1472-6750-7-16
- Machine Learning
- Amino Acid Substitution
- Active Variant
- Partial Less Square Regression
- Machine Learning Algorithm