The Trans-Proteomic Pipeline (TPP) is a suite of software tools for the analysis of MS/MS data sets. The tools encompass most of the steps in a proteomic data analysis workflow in a single, integrated software system. Specifically, the TPP supports all steps from spectrometer output file conversion to protein-level statistical validation, including quantification by stable isotope ratios. We describe here the full workflow of the TPP and the tools therein, along with an example on a sample data set, demonstrating that the setup and use of the tools are straightforward and well supported and do not require specialized informatic resources or knowledge.
MS/MS has enabled the identification of large numbers of proteins in biological samples in a high throughput manner 1, in an approach termed shotgun proteomics. Peptides from digested proteins are chromatographically separated and then fragmented in the mass spectrometer, yielding fragment ion spectra that can be used to deduce the corresponding peptide sequences. These in turn allow the identification of proteins in the sample. In addition to the protein identities, quantitative information can also be obtained through a variety of stable isotope and isobaric tagging peptide labeling techniques 2. MS/MS has therefore become the most commonly used method in proteomics today.
There are a wide array of workflows, mass spectrometers, and data analysis tools available. There are typically multiple steps in the data analysis process, with a variety of techniques and options in the software tools used for each step 3. A typical modern workflow for the analysis of MS/MS data that includes validation and quantification can be quite complex, and is described by Deutsch et al. 4: conversion of raw vendor files to an open (non-proprietary) format that all tools can access downstream; identification of spectra with a sequence or spectrum library search engine; statistical validation of the putative identifications; quantification of results if a quantitative labeling strategy was used; and finally protein inference and interpretation.
Many tools are available from different groups and can be strung together in custom ways, but there are only a few suites of tools that aim to provide a single environment for performing all or most steps in the workflow, including the openMS proteomics pipeline (TOPP) 5, MaxQuant 6, and the Trans-Proteomic Pipeline (TPP) 7. The TPP, developed at the Seattle Proteome Center (SPC), is the oldest and most comprehensive fully open-source suite of software tools that facilitates and standardizes the analysis of LC-MS/MS data. The TPP includes software tools for MS data representation, MS data visualization, peptide identification and validation, quantification, and protein inference. In this study, we provide a review of the available TPP tools along with a tutorial demonstrating how easy the analysis and validation of a data set with the TPP can be.
2 Online tutorial
An important aspect of the effectiveness of a set of tools is their usability. A tool that implements a clever algorithm but is very difficult for the average researcher to use, or is not linked or linkable to tools performing upstream or downstream steps, is of only modest value. Similarly, a tool for which the underlying algorithm is not described is difficult to assess. Significant effort was, therefore, devoted to make the TPP and all of its tools easy to use and well integrated into a robust system. To illustrate this, we have developed a short tutorial of ten steps to accompany this article that demonstrates the use of the tools from installation to final analysis of a sample data set. The tutorial can be found at http://tools.proteomecenter.org/wiki/index.php?title=TPP_Demo2009.
3 TPP analysis overview
The full workflow of a typical MS/MS proteomics dataset through the TPP is summarized in Fig. 1. Raw mass spectrometer output files are first converted to an open format such as mzXML 8 or mzML 9. These files are run through one or more search engines such as X!Tandem 10, Mascot 11, SEQUEST 12, or SpectraST 13, and the results are converted to the pepXML 7 file format. The Pep3D tool 14 can be used to visually inspect each MS run to assess the quality of the chromatography. PeptideProphet 15 is then used to validate the search engine results and to model correct versus incorrect peptide–spectrum matches (PSMs). The data sets can then be validated at the peptide-identification level with the new iProphet tool (Shteynberg, D., Deutsch, E. N., Lam, H., Eng, J. H. et al., Postprocessing and validation of tandem MS datasets improved by iProphet; in preparation). Finally, protein-level validation and protein inference are performed with ProteinProphet 16. Within this workflow, the quantification analysis tools XPRESS 17, ASAPRatio 18, or Libra 19 may be used with data that derive from isotopically or isobarically labeled samples. The final output is a protXML 7 file that includes all the protein- and peptide-level information, including the finally assigned probabilities for all peptides and proteins. This process will be described in more detail in the following sections, along with references to the aforementioned tutorial.
3.1 Sample dataset
The sample dataset used in this tutorial is derived from a SILAC 20 metabolically labeled yeast (Saccharomyces cerevisiae) whole-cell lysate that was digested with trypsin and run on a Thermo Scientific LTQ Orbitrap mass spectrometer as described in detail in the Supporting Information. The raw and processed data are available in the PeptideAtlas 21 raw data repository as PAe001337. Downloading of the sample data set is covered in step 2 of the tutorial, after the installation of the TPP, described in Section 3.2.
Considerable effort has been expended to make the TPP easy to install. The TPP can be installed on most major platforms, including Microsoft Windows, UNIX/Linux, and MacOS X. The installation is the easiest on the Windows platform, and as this is the most commonly used environment, we have tailored the tutorial around a session on the Windows platform. However, except for installation, all of the remaining steps will be nearly identical under other platforms. It should be noted that the primary development environment and the use of the TPP by the authors is under Linux, and therefore, the support for UNIX-like platforms is truly native rather than merely an afterthought.
Step 1 in the tutorial is the preparation for using the TPP on a computer. Here, the installation of the TPP is described as well as launching the graphical user interface. Installation under Windows takes 10–30 min, after which the software is ready to use. Installation on Linux or MacOS X platforms does involve some more effort. An installation guide for each of these platforms is available as described in the tutorial.
The TPP is a collection of over 30 tools (Table 1) that can be strung together as a pipeline or run individually as needed. A graphical user interface, Petunia, provides a point-and-click environment from within an ordinary web browser, allowing users to launch tools separately or launch a complete analysis from a single page.
Table 1. List of tools in the Trans-Proteomic Pipeline distribution
a) Requires vendor DLLs.
b) Third-party software included in TPP.
Graphical user interface for TPP and most other tools in this table.
Common, open-source search tool. Augmented with K-score algorithm, optimized to work with PeptideProphet.
Sequence search tool
Search tool developed at SPC.
Sequence search support
Creates reverse databases for decoy searching.
Spectral library search tool
Identifies MS/MS spectra by comparing against a spectral library.
Spectral library support
Converts a spectral library to HTML format
Converts SEQUEST results to pepXML.
Converts MASCOT results to pepXML.
Converts X!Tandem results to pepXML.
Provides statistically rigorous probabilities for search identifications.
Statistical modeling: internal tool
Maps peptides in a pepXML file to a protein sequence database different from the search database. Needed for spectral library searching.
Further refines PeptideProphet probabilities. Can combine results from multiple searches.
Provides statistically rigorous probabilities for protein identifications.
Statistical modeling: post-processing
Summarizes TPP results for an experiment. Only available via command line, not via Petunia.
Viewer for pepXML files.
Viewer for protXML files.
Command line interface
Runs TPP statistical modeling tools from command line.
Finds high quality spectra among those unassigned by search.
Quantitative analysis for isotopic labeling (simple tool).
Quantitative analysis for isotopic labeling (more sophisticated tool).
Quantitative analysis for multichannel isobaric labeling.
Creates SRM transitions lists from a set of spectra.
3.3 Conversion to an open format
Each instrument vendor has a different proprietary file format for its mass spectrometer output files; in fact, some vendors have several formats depending on the instrument type. In order to support all kinds of data from all mass spectrometers, the TPP allows the conversion of the data into either of two open, vendor-neutral formats: mzXML or mzML. The mzXML format was developed by our group. Recently, this format has been supplanted by the mzML format, developed by the HUPO Proteomics Standards Initiative (PSI). Both mzXML and mzML are currently supported by TPP.
The TPP tools used for the conversion of file formats include msconvert 22 (as well as the older ReAdW tool) for Thermo Scientific RAW files, mzWiff for ABI/Sciex WIFF files, Trapper for Agilent files, and massWolf for Waters' MassLynx files. Bruker provides a freely downloadable tool, CompassXport, for the conversion of their formats to mzXML. All of these conversion tools can only work on a Windows computer with vendor libraries installed. However, some vendors are starting to agree to allow the distribution of these libraries free of charge. The current status of conversion tools can be found in http://tools.proteomecenter.org/wiki/index.php?title=Formats:mzXML.
To perform the conversion, follow the detailed items in step 3 of the tutorial, which demonstrates how the vendor RAW files for the sample data set are transformed into mzML through Petunia using msconvert. If a computer does not have the appropriate Thermo Scientific libraries installed, the tutorial shows how to obtain pre-converted mzML data files for use within the tutorial.
3.4 Search with a sequence search engine
The main step in processing MS/MS data is a search with a spectrum interpretation algorithm that attempts to match the spectra with possible peptide sequences extracted from a given list of protein sequences. This is most commonly performed with one of the many available sequence search engines. It is a common misconception that the TPP only supports SEQUEST search results, as it did at the very beginning. In reality, the TPP now supports X!Tandem, Mascot, SEQUEST, OMSSA 23, Phenyx 24, and ProbID 25. Only X!Tandem, an open source search engine, is bundled with the TPP; other engines must be installed separately.
In the tutorial step 4, we demonstrate the search of the sample data set with the bundled X!Tandem. For better results, the X!Tandem packaged with TPP was augmented with a pluggable scoring mechanism 26 called K-score, which has been optimized to work well with PeptideProphet.
3.5 Search with a spectral library search engine
An alternative or complementary approach to sequence searching is spectral library searching. In this approach, spectra are compared against a library of previously identified spectra rather than against hypothetical spectra predicted from a sequence database. Spectral library searching is shown to be much faster and more capable of identifying low-quality spectra than sequence search engines 13. This is mainly because spectral searching benefits from a smaller search space (fewer candidates to choose from) and the use of real reference spectra as opposed to theoretical ones predicted, often simplistically, by sequence search engines. The obvious limitation is that only peptides corresponding to spectra in the library can be identified; previously undetected peptides will be missed. Therefore, spectral searching is best seen as a complementary approach to sequence searching, and the best results are often obtained from combining results from multiple search engines, as described in Section 3.8.
In step 5 of the tutorial document, we demonstrate how to download the yeast ion trap spectrum library version 2 created by the National Institute for Standards and Technology and search the sample data set using the SpectraST 13 spectral library search engine. SpectraST is part of the TPP and every TPP installation already comes with SpectraST.
3.6 Validating search results with PeptideProphet
Every search engine returns one or more best interpretations (PSMs) of each input MS/MS spectrum along with scores that quantify the quality of the matches. Some engines provide a normalized score (e.g. an E-value) that indicates the likelihood that the PSM is an incorrect, random event, whereas other search engines provide raw scores without absolute significance. It is thus challenging to directly compare the results of different search engines. The PeptideProphet algorithm 15 was developed to separate all of the resulting PSM scores into a population of correct and incorrect PSMs using an expectation maximization algorithm. PeptideProphet assigns a probability of correctness to each PSM and calculates a set of global false discovery rates 27 as a function of the probability cutoff. In step 6 of the tutorial, this statistically rigorous validation of both the X!Tandem and the SpectraST search results is demonstrated.
In calculating PSM probabilities, PeptideProphet can be made to consider not only the search score, but also PSM attributes such as mass deviation, number of enzymatic termini, number of missed cleavages, and retention time. This allows better discrimination between correct and incorrect interpretations and leads to improved probability estimates. PSMs whose attributes conform better to those of the correct population are rewarded, whereas PSMs whose attributes conform better to those of the incorrect population are penalized. For example, a PSM whose mass deviation (between that measured by the instrument and that predicted if the sequence interpretation is correct) is near the average for other correct identifications (not necessarily 0.0) is rewarded with a higher probability.
When the dataset is from a high accuracy instrument, PeptideProphet can employ a high mass accuracy model which models the deviation in observed mass to the nearest isotopic peak of the putative identification. The overall accuracy is learned from the data and, therefore, the better the accuracy of the input masses, the better the discrimination by mass deviation. If external tools such as DeconMSn 28 are used to improve the input precursor m/z values, PeptideProphet learns this and applies a more discriminating model. An additional recently developed model is the elution time model, which fits for an entire MS run the actual elution time for each PSM to its predicted elution time. In the same way, PSMs that conform well to the predicted model are rewarded and large deviations are penalized. A study describing these features in detail is in preparation.
PeptideProphet does not require a sequence search performed with a target-decoy strategy 29 as input in order to perform its modeling. An ordinary target-only search can significantly decrease computation time over target-decoy strategies. However, if a target-decoy strategy is employed, PeptideProphet can use this information to help refine the models 30, or even model the incorrect population directly to the decoy population 31. PeptideProphet can also be set to ignore decoy information during modeling so that the decoy PSMs can be used for independent evaluation of the TPP results.
The final result of PeptideProphet analysis is a pepXML format file that reports a probability for every PSM as well as the results of the modeling, including a representation of the receiver operating characteristic curves. The pepXML file can be interactively explored with PepXMLViewer application.
3.7 Visualizing the chromatography with Pep3D
There are a number of problems with chromatography that can degrade the number of identifiable MS/MS spectra in a run. For example, non-optimal chromatography can result in the sample eluting over a short range of the entire LC gradient or polymer contamination can wastefully consume CID scan acquisitions. Most of these problems are not apparent by looking through the search results. However, a visual image derived from all the MS1 spectra can make apparent many chromatography problems. The TPP tool Pep3D 14 can be used to create images from one or more LC runs for quality control and visual inspection. Overlaid on the images are symbols for each MS/MS spectrum acquisition, color coded when the MS/MS spectrum is identified with high confidence. Figure 2 shows the Pep3D output for one of the sample runs. Step 7 in the tutorial demonstrates the use of Pep3D to examine the sample runs.
3.8 Further peptide-level validation with iProphet
After validation of a dataset with PeptideProphet, each PSM is assigned a probability of being correct. However, there is still a significant amount of corroborating evidence from other identifications that can be used to further discriminate between correct and incorrect PSMs. The iProphet tool (Shteynberg, D. et al., in preparation) is an additional tool that can improve the probability estimates coming out of PeptideProphet. In addition, it offers the opportunity to combine search results from multiple search engines. In step 8 of the tutorial, iProphet is used to combine the results of the X!Tandem and SpectraST searches of the sample dataset.
When a dataset is searched with two or more search engines, a mostly overlapping yet somewhat different set of spectra are scored highly by PeptideProphet. When a spectrum is scored highly by multiple search engines, it is more likely to be correct (although, of course, multiple search engines can very occasionally agree on the wrong answer). If two or more search engines disagree on the matching peptide sequence, each probability is penalized by the magnitude of the probabilities of the conflicting results.
Besides corroborating evidences from multiple search engines, iProphet also takes into account repeated discoveries of the same peptide ion, or the same peptide sequence with a different charge state or modification, all of which increase confidence in the identification. The iProphet tool creates mixture models for several lines of corroborating evidence (number of sibling search results, number of replicate spectra, number of sibling ions, number of sibling modifications, and number of sibling experiments) and adjusts the models for each dataset processed. For each model, the probability of each PSM is either rewarded or penalized based on its similarity to the correct population or incorrect population, respectively. The final result is better discrimination, yet still with accurate probabilities and global false discovery rate.
3.9 Differential labeling quantification analysis
Isotopic labeling is an important technique for adding a quantitative dimension to shotgun proteomics experiments. In techniques such as ICAT 32 or SILAC 20, one or more samples are labeled with heavy isotopes and then run through an instrument together with matched samples consisting of light isotopes (control versus treatment(s), disease versus normal, etc.). The relative abundances of two samples can be measured via the ratio of the two extracted ion currents of a heavy-light peptide pair. There are two TPP tools for the analysis of such data. XPRESS 17 is the earlier, simpler tool that is still appropriate for some uses. The more recent ASAPRatio 18 is more sophisticated in its measurement of, and aggregation of measurements from, multiple peptide ions from the same peptide, as well as aggregation at the protein level. Step 9 of the tutorial demonstrates the use of ASAPRatio on the sample data set to derive abundance ratios for the two samples.
Another popular technique is isobaric labeling such as iTRAQ 33 and TMT 34, where peptides in up to eight samples can be labeled with N-terminal isobaric labels, each with a different fragment ion peak near 100 m/z. Measurement of the relative peak intensities of these reporter peaks in the fragment ion spectra yield relative abundances of peptides and hence proteins in the original samples. The TPP tool Libra performs the quantitative analysis of such data sets, although this is not illustrated in the tutorial. Although TPP tools can be used to aid with quantification on label-free MS/MS data sets using extracted ion chromatogram or spectral counting techniques, these workflows are not explicitly supported.
3.10 Protein inference and validation with ProteinProphet
The final step of processing data with the TPP entails combining all of the peptide observations into a final list of proteins. This is a complex problem, mainly due to the fact that many related proteins share peptide sequences 16. The ProteinProphet tool 35 performs this final analysis step.
ProteinProphet first applies a mixture model based on the number of distinct peptides per protein (sibling peptides) to boost the probabilities of peptides with multiple siblings while penalizing peptides without siblings. ProteinProphet then performs a protein inference analysis to create the simplest list of proteins that can explain all the peptide observations. Each protein is assigned a probability of being in the sample. Step 10 in the tutorial demonstrates the analysis of the sample data set with ProteinProphet.
The final result is a protXML 7 file with a list of all proteins corresponding to the PSMs, along with protein probabilities and global false discovery rates at different thresholds. Each protein is annotated with the relative abundance ratio, if a quantification technique was used, along with uncertainties. The output can be viewed using the protXML viewer of the TPP, and exported to Excel or other analysis tools in order to address the biological question that prompted the experiment. A few of these tools are described in the following sections.
3.10.1 Data formats
In order to support the vision of the TPP as a complete set of tools to analyze shotgun proteomics data from any instrument or vendor, it was necessary to develop common, open (non-proprietary) formats through which the tools can exchange data. As there were no such formats at the time, we developed and made accessible the mzXML, pepXML, and protXML formats. The mzXML is used for raw spectra, as described in Section 3.3. Search engine results are converted to the pepXML format for downstream analysis. PeptideProphet reads pepXML files from the search engines and creates a combined pepXML file with all the model information. The iProphet tool also reads pepXML and writes out new pepXML with additional model information. The quantification tools write their results into the pepXML files. Finally, ProteinProphet reads one or more pepXML files and writes out a protXML file that is the final end result of TPP processing, including protein identifications and supporting information. None of these formats is an official standard, but all have become de facto standard formats in the community and are used by many software tools unrelated to the TPP, primarily due to the free availability of software that reads and writes these formats.
The TPP team worked with the PSI MS standards working group to develop a next-generation format for mass spectrometer output files, mzML 9, that meets not only the needs of the TPP, but also other needs in the community. The PSI proteomics informatic group, also with cooperation from the TPP development team, has developed a new format that can encode downstream informatic analysis of proteomics MS data. The new format, mzIdentML, combines the information encoded in pepXML and protXML and much more into a single file format (except for quantification information, which is expected from a subsequently released mzQuantML format). Although it is likely that the TPP will continue to use pepXML and protXML as internal working formats for the pipeline, in the future the TPP will convert all the final results to mzIndentML once its development is complete.
All of the information required by the Guidelines for Proteomic Data Publication 36, 37 is encoded in the standard file formats mzML, pepXML, or protXML, with the exception of the annotated spectra. Therefore, these files can be included with journal submissions to fulfill the requirements. For the annotated spectra, we have developed a program that will generate a PDF document of annotated spectra as required by the guidelines. The functionality is currently in beta testing (beta version available in the source code repository) and will be officially released in version 4.4 of the TPP soon.
3.10.2 Subsequent analysis
The final result of a TPP analysis is a protXML file containing a list of proteins, their probabilities of presence in the sample, supporting peptide information, and possibly quantification results. The results can be browsed with the protXML viewer or exported into a tab-delimited format suitable for Excel. However, many other tools exist for further downstream analysis.
The Protein Information and Property Explorer 38 is a web-based interactive interface developed at the SPC for exploring the significance of protein lists. It includes functionality for mapping to other identifiers (Entrez gene, Unigene, Uniprot, etc.), gene ontology enrichment calculations (to see which biological processes are overrepresented in the list), annotation of the list to specific gene ontology terms, and more. The Protein Information and Property Explorer can also serve as a data management tool, saving lists with user specified metadata (i.e. description of the data).
The results of TPP processing can be imported into the Systems Biology Experiment Analysis Management System (SBEAMS) Proteomics database, which is part of the SBEAMS Project 39, a framework for collecting, storing, and accessing data produced by proteomics experiments and other experiment types. Several other database systems such as CPAS 40 and YPED 41 provide a data management layer on top of the embedded TPP tools.
3.10.3 Other tools
In addition to the tools describe herein, many other tools come bundled with the TPP. The full set available is summarized in Table 1. A few tools of special note are described hereafter.
The xinteract tool is a command-line wrapper program that can be used to run several of the TPP tools at once. It has an extensive set of parameters that can be used to chain together programs such as PeptideProphet, iProphet, and ProteinProphet in one command. This is often quite useful in an environment where the processing of many experiments is scripted or for other automation scenarios.
The QualScore 42 tool takes a PeptideProphet analysis result as input and takes very high probability PSMs and very low probability PSMs as proxies for high quality and low quality spectra, respectively. Then, a machine learning technique attempts to discriminate between high quality and low quality spectra based on the attributes of the input training sets. The result is then a population of spectra that share many attributes with the population of highest probably PSMs, but could not be identified. This population is suitable for another round of exhaustive searching with expanded modification parameters or protein lists.
The MaRiMba tool 43 is a workflow for selecting selected reaction monitoring (SRM; sometimes called multiple reaction monitoring or MRM) transitions for a targeted proteomics assay based on a consensus spectral library as created by SpectraST. MaRiMba creates SRM transition lists from downloaded or custom-built spectral libraries, restricts output to specified proteins or peptides of interest, and filters SRM lists based on user-defined precursor peptide and product ion properties. MaRiMba can also create SRM lists containing isotopically heavy transitions for use with isotopic labeling strategies such as SILAC. MaRiMba outputs the final SRM list to a text file convenient for upload to a mass spectrometer. The TPP itself does not presently provide analysis for acquired SRM data; however, the TIQAM (Targeted Identification for Quantitative Analysis by MRM) 44 and MRMer 45 programs are available and the ATAQS (Automated and Targeted Analysis with Quantitative SRM) pipeline is under development and will be distributed by our group similarly to the TPP.
4 Support for the tools
Support for the tools is provided via an email discussion list at Google Groups, with a browsable archive. Over 800 users are signed up for the list and many questions are answered by members of the TPP community other than those at the SPC. Members of the TPP team are often found in a booth at major conferences helping to install the TPP on laptops of booth visitors or answering questions. The SPC also hosts a 5-day course twice a year. At each course, 30 students are accepted to attend lectures describing each of the tools, usually taught by the original authors or active maintainers of the software. Students work in pairs on the provided course computers to process sample data sets (including the data set described in the accompanying on-line tutorial) through the various tools in guided tutorials and independent exercises. All information about the email discussion list, booths, and courses are available on the SPC web site.
As mentioned in Section 1, there are several software packages similar to the TPP that are freely available to the community. OpenMS provides a large number of C++ classes that can be used to facilitate the building of proteomics software tools, and TOPP (built on OpenMS) consists of several small applications that can be put together to create a custom analysis pipeline tailored for a specific problem. Mascot and OMSSA are supported and a tool is available to create a consensus of two searches of the same data. TOPP uses the PSI mzML format as input and then implements additional XML-based formats for internal communication among tools, similar to pepXML and protXML. TOPP is available for all three major operating system platforms and is completely open source, but lacks the more sophisticated modeling algorithms and quantification tools of the TPP.
MaxQuant is for use with high-resolution LTQ FT and LTQ Orbitrap instruments from Thermo Fisher. This software uses RAW files directly and, therefore, standard formats are not needed or supported. The software works on Windows XP or Vista, is used with the MASCOT search engine, and can perform quantitative analysis with the SILAC technique. For experiments that fall within this scope, MaxQuant provides an effective analysis platform. It is freely available as a downloadable binary, but the source code is not open.
Among free software solutions, the TPP provides the most advanced analysis algorithms, and the only ones that are truly able to improve upon the native results for a variety of search engines from data sets derived from a variety of instrument vendors and types. However, the TPP is not without shortcomings. Installation difficulties still provide an immediate barrier to some, although installation is greatly improved over earlier versions. The user interface is sometimes lacking; considerably more effort has gone into developing advanced algorithms than advanced user interfaces. Although the algorithms work well on very large data sets, there are known deficiencies with the data viewers that cause performance problems with large data sets. The large numbers of configurable options can be daunting, although the default values are usually adequate. No features are available to assist with annotation of experiments, e.g. sample information. Finally, converting the raw vendor formats can be problematic for some vendors, and if the raw files cannot be converted to mzML or mzXML, then the TPP cannot be fully applied. Work is underway to address these shortcomings, although solutions will take time.
6 Future outlook for the TPP
The TPP is a complete and working system that forms part of the routine analysis workflow of hundreds of researchers. Yet robust development activity by the TPP team continues. The tools are continually adapted to improvements in wet laboratory techniques and advances in instrumentation. The iProphet tool has just recently been completed. Adaption of the tools for instruments that can use the novel fragmentation technique electron transfer dissociation has recently been completed 46, including support for higher charge states, additional enzymes, charge-state prediction algorithms, electron transfer dissociation spectrum library building, and improved annotation of c- and z-type ions.
We have preliminary beta support for some additional search engines such as InsPecT 47 and MyriMatch 48 with support for more search engines planned, as well as support for additional quantitative labeling techniques. Work is underway to adapt the TPP to work seamlessly in a cloud-computing environment. We envision a system whereby a user can instantiate a preconfigured node on the Amazon Elastic Compute Cloud (EC2) already running the TPP (by paying a modest usage fee to EC2) as well as instantiating a user-selectable number of other nodes to work as search engine slave nodes for processing large jobs.
The TPP will be better integrated with the current proteomics data repositories. From within the TPP interface, users will be able to easily upload their results to one of the repositories such as PeptideAtlas 21, PRIDE 49, Tranche 50, or Peptidome (http://www.ncbi.nlm.nih.gov/projects/peptidome/). Similarly, the interface will be adapted to easily download raw data from Tranche or the other repositories for viewing or reprocessing in the TPP environment. We are also planning a visual dashboard application that provides a quick overview of the results of TPP processing with auto-generated charts and figures, as well as an analysis advisor that can point out common problems in processing, sample preparation, or instrument methods. These and other developments will continue to provide an ever better user experience and a more capable suite of tools to keep up with improving MS/MS proteomics techniques.
7 Concluding remarks
We have described the various tools that comprise the TPP, and demonstrated the utility of the TPP via a sample data set in the accompanying tutorial. The TPP provides both a graphical user interface for ease of use in addition to a discrete set of back-end tools that may be easily combined into custom pipelines for large amounts of data, as is used to enable the PeptideAtlas project. Further, the TPP tools provide an advanced, statistically sound, and continually improving suite that is a huge improvement upon the use of a single search tool with no post-validation. For users who want an intensive hands-on training in the use of these tools, the Seattle Proteome Center offers a biannual course that has trained over 500 members of the community thus far.
The TPP tools are all free and open source, and have become more robust and far easier to use in the past few years. For the Windows platform, there is an easy-to-follow quick installation guide available at the Seattle Proteome Center web site (http://www.proteomecenter.org). The TPP is also compatible with Linux and MacOS X, although the installation for those platforms is somewhat more complex. However, the entire process is sufficiently simple that all users can easily take advantage of the advanced tools in the TPP to analyze their data.
The authors are grateful to the many users of and contributors to the TPP project for their feedback, bug reports, and code contributions. This work has been funded in part with federal funds by the National Heart, Lung, and Blood Institute, National Institutes of Health, under contract no. N01-HV-28179 and grant R44HG004537, and from PM50 GMO76547/Center for Systems Biology.
The authors have declared no conflict of interest.