当前位置: 首页 > 期刊 > 《生理学报》 > 2005年第4期
编号:11326671
Broad-based proteomic strategies: a practical guide to proteomics and functional screening
http://www.100md.com 《生理学报》 2005年第4期
     Departments of Medicine, Division of Cardiology Pharmacology and Molecular Sciences Biomedical Engineering Biological Chemistry, Johns Hopkins University, Baltimore, MD, USA

    Abstract

    Proteomics, the study of the proteome (the collection of all the proteins expressed from the genome in all isoforms, polymorphisms and post-translational modifications), is a rapidly developing field in which there are numerous new and often expensive technologies, making it imperative to use the most appropriate technology for the biological system and hypothesis being addressed. This review provides some guidelines on approaching a broad-based proteomics project, including strategies on refining hypotheses, choosing models and proteomic approaches with an emphasis on aspects of sample complexity (including abundance and protein characteristics), and separation technologies and their respective strengths and weaknesses. Finally, issues related to quantification, mass spectrometry and informatics strategies are discussed. The goal of this review is therefore twofold: the first section provides a brief outline of proteomic technologies, specifically with respect to their applications to broad-based proteomic approaches, and the second part provides more details about the application of these technologies in typical scenarios dealing with physiological and pathological processes. Proteomics at its best is the integration of carefully planned research and complementary techniques with the advantages of powerful discovery technologies that has the potential to make substantial contributions to the understanding of disease and disease processes.
, http://www.100md.com
    Introduction

    Physiology, and scientific research in general, has tended towards a reductionist approach in understanding diseases and the disease processes. Traditionally this strategy has been necessary due to the limitations of most technologies, the complex nature of physiological and pathological systems and the successes related to drilling deeply into a focused area eliminating confounding variables. However, new proteomic technologies have created a renaissance for the classical approaches that deal with understanding complex systems and diseases at a global level, thus allowing an expanded experimental view. Proteomic approaches were initially explored with the advent of two-dimensional gel electrophoresis (O'Farrell, 1975) and application of chromatography techniques (Neverova & Van Eyk 2002); however, the difficulty in identifying proteins suppressed the broad application of proteomics. This technological hurdle has for the most part been overcome through the application of mass spectrometry (MS) used in combination with developing genomics databases. The goal of this review is twofold: the first section provides a brief outline of proteomic technologies, specifically with respect to their applications to broad-based proteomic approaches, and the second part provides more details about how to apply these technologies in typical scenarios dealing with physiological and pathological processes.
, 百拇医药
    Proteomics defined in its most broad terms is the study of the proteome: the collection of all the proteins expressed from the genome in all isoforms, polymorphisms and post-translational modifications (PTMs). Broad-based proteomics is a strategy wherein disease or physiological models are analysed with the most well-suited proteomics technologies to reveal changes or differences in protein make-up between the experimental condition and controls. This approach allows for the optimal discovery of the changes which define the model system. Screening for the maximum number of protein changes limits experimental bias by not excluding groups of proteins, therefore making it imperative to use the most appropriate technology for the biological system and hypothesis being addressed. The initial proteomic discovery is only the beginning of such a study and must be complemented using a variety of laboratory-based techniques to put the protein changes into their cellular context, as outlined in Fig. 1. This multitiered approach allows us to ask global questions arising from an overview of the disease or physiological condition, and is particularly powerful when applied to studies of protein function by knockout, knock-in or transgenic expression in animals like the mouse. For example, identification of the primary protein target and the composition of related downstream targets; whether the changes involve PTM or expression differences and how these two parameters may be regulated; the timeline for the response (do the protein changes occur in parallel or sequentially); and the relationships between protein synthesis and degradation of the target – to name a few. Thus, initial data obtained by broad-based proteomics are the framework upon which downstream experimental decisions are based. In the second tier, the initial proteomic screen will pinpoint which subproteome or approach to use in the next level of analysis, or whether to look at multiple different disease models to determine the converging responses. Occasionally, very direct conclusions can be drawn by the results that reveal the changes that are unique from one condition. For instance, in cardiac disease a condition like hypoxia may be differentiated from another condition like ischaemia or in general terms a diseased tissue may be differentiated from a control tissue.
, http://www.100md.com
    Technologies and considerations

    Create a general hypothesis specific to the study system

    The key to increasing the chances for success in any proteomics experiment is to have a thorough understanding of the physiological model or disease process being studied, from which a hypothesis is formed that will drive the choice and selection of the particular proteomic/analytical approach. For example, the proteomic approach will differ for studies in which the pathological perturbation changes the high copy number (high abundance) proteins which dominate a cellular proteome (cytoskeletal proteins, many housekeeping proteins, etc.), versus those that are confined to low copy number – low abundance proteins (e.g. chaperones, signal transduction molecules, etc.). Understanding the experimental system will greatly improve the chances of success on proceeding to the next phase of a proteomics project, or at least determine if proteomics is an appropriate experimental design for the goals.
, 百拇医药
    Choice of model

    Proteomics is built upon the foundations of genomics. Therefore, a lack of genomic information on a particular species can substantially limit the success in a proteomics project by increasing the difficulty of successful protein identification. Because of this, the first question a physiologist must address is the choice of model system in which to attempt to apply proteomics, and whether adequate genome coverage exists, while balancing the appropriateness of the model for the specific biological question. Although genome projects involving many of the traditional species used as classical physiological models are now under way (e.g. rabbit, dog and sheep), the current lack of a genome sequence for a particular species, especially animals classically used in many in vivo studies, can be a tremendous stumbling block for a proteomics project, since even having proteins with 80% sequence identity is often insufficient for protein identification through homology by peptide mass fingerprinting (PMF) (Wilkins & Williams 1997) or MS (it is only with de novo sequencing that protein or genome databases are not relied upon for homology; Steen & Mann 2004). Therefore, until the genome of the model species of interest is fully sequenced, the first recommendation is to work with a model where a full genome is in place, or to choose a collaborator that has experience managing projects involving these species. Animal models, however, are of extreme utility in proteomics studies as they eliminate the heterogeneity often observed when using human tissue and allow careful control of temporal and perturbation studies.
, 百拇医药
    Choice of proteomic approach

    With broad-based proteomic approaches one is generally interested in maximizing coverage of the proteome or subproteome (a subfractionated or enriched subset of the proteome) under study. Often termed protein expression profiling, the relative changes between proteomes (or subproteomes) are examined either quantitatively (protein quantity) or qualitatively (determination of PTMs and isoforms present). It is important to realize that not all proteomics technologies provide both qualitative and quantitative data – thus the choice of technologies will be dictated by the importance of these two kinds of data in the system under study (Table 1). For instance, in acute injury the immediate cellular response will primarily be changes in the PTM status. General mining (brute-force identification of proteins) to identify all the proteins in a particular sample, and interaction mapping are more specialized applications that require either a high-throughput proteomics laboratory or a facility utilizing yeast genetics as a starting point for identifying potential interaction partners, although most laboratories are capable of using this approach in a limited manner with affinity purification strategies. Today, there is a general lack of PTM characterization. However, there are several approaches including affinity approaches, and the use of new and specific dyes that enrich the detection of specific PTMs (Peng et al. 2003; Ge et al. 2004). Certainly, in the future development of robust methods for detection of specific PTMs and the identification of the exact amino acid residue(s) that are modified will result in an explosion of information.
, http://www.100md.com
    Sample complexity

    Prior to the initiation of any proteomic project, some basic attributes of the sample being analysed must be considered. Broadly these are protein concentration, dynamic range (highest to lowest abundant proteins), degree of protein solubility and the copy number of the protein classes in which one is attempting to assess changes. Despite the femtomolar sensitivity of modern mass spectrometers, issues involved in separating complex protein mixtures (protein interactions, etc.) and obtaining good quality MS data (residual contaminants, etc.) often raise the sample requirements for biological samples, both from an aspect of protein separation and MS identification. A rule of thumb is to expect reasonable success in identifying your protein of interest if the protein can be visualized by coomassie blue staining (> 300 ng), and with developed skill and good instrumentation, success can be obtained in identifying high abundant proteins that are visible by silver staining (> 5 ng) (Shevchenko et al. 1996). Combining samples for multiple preparations or from preparative analysis will increase the likelihood of protein identification.
, 百拇医药
    Applications to disease screening

    Choice of sample

    Figure 2 illustrates the decision schema for assessing the feasibility of a proteomic project. The first consideration as outlined above is identifying the complexity of the sample, and the source of the sample. Typical targets for sample acquisition from humans are blood (serum/plasma/cells), cerebrospinal fluid (CSF), urine, mucosal secretions and tissue (biopsy/postmortem), where ease of collection is generally the driving motive, however, quantities may be limited. Ideally, samples from humans are obtained from individuals over time and cohorts have to be carefully controlled to minimize any variability within sample groups. Animal models (depending on size and cost) generally allow researchers to obtain more abundant samples and specifically control where a sample is obtained. In some cases, an appropriate cell culture model exists, where the amount of material can be expanded to suit the large quantity of protein needed for a project. It is necessary to guard against the frustrating position in which the quantity of material is only sufficient to identify the number of proteins changing, and it is impossible to obtain more of the same sample for follow-up experiments such as the actual protein identification, characterization of isoform/PTM, or functional analysis. To avoid this, it is best to have abundant sample (preferably an expendable control sample), where multiple experiments optimizing conditions can be performed, prior to the use of precious samples. As mentioned above, the complexity of the sample must be addressed, as illustrated by the triangle in Fig. 2 in which the interplay between complexity, extraction conditions and reproducibility is shown. Basically, the simplest most reproducible approach should be utilized to obtain the proteome of interest, and if necessary, the same rule should apply to obtaining the subproteome of interest. Generally, one should target the source organ or tissue directly, sampling diseased tissue with an internal control of non-diseased tissue if possible, or at least an exactly age and condition matched control.
, http://www.100md.com
    This schema is intended to guide the researcher to the most beneficial (most accurate, appropriately scoped, least-biased) analysis of the sample. The gradient informatics box highlights the majority of the proteomic data analysis and often is the bottleneck in proteomic analysis.

    Separation technologies

    The choice of protein separation tool used is critical as the protein/peptide separation technology must match the experimental hypothesis. Table 1 illustrates some of the potential tools at the disposal of a typical investigator. Protein separation technologies exploit the intrinsic properties of proteins (molecular mass, isoelectric point (charge), hydrophobicity and biospecificity (or immuno-specificity)).
, 百拇医药
    System familiarity. Due to the complexity of any proteomic study the investigator needs to be familiar with the technical aspects even when done in collaboration with a core laboratory. Care must be taken when using technologies and bioinformatics that an investigator does not have direct control over or is not able to directly participate in the process as there are many places where judgement is required. Proteomics is not yet (and most likely will never be) a cook-book approach. There are still many issues that confound results, from protein solubility issues and issues in chromatography (elution profiling) to deconvolution of large data sets, MS coverage, automated analysis settings and bioinformatics analysis choice. Unless an investigator has direct access to the raw data for analysis, and a level of understanding where they can manipulate the raw data, it is unwise to present the data as a discovery; rather it is best to confirm the results by classical biochemical technique. The bottom-line ‘keep it simple’ strategy is most often the best choice for most investigators in regard to proteomics unless they are either making a commitment to proteomics or in collaboration with proteomic investigators.
, 百拇医药
    Protein abundance/protein characteristics/sample complexity. The choice of separating proteins or their peptide complements (obtained from a globally digestion of the proteome) and the separation technology must be based foremost on the abundance and the character of the proteins suspected to change in the experiment. For instance, if PTMs are thought to dominate the proteomic changes, then the separation of intact proteins is often needed. Additionally, care must be taken if the goal is the detection of low copy number proteins as they will be under-represented by most techniques. Digestion of the proteome and analysis of the resulting peptide fragments by liquid chromatography (Table 1) can often enhance the detection of lower abundance proteins, but this can be at the cost of losing information about the nature of the proteins (isoform and PTM status). Specific enhancement of low abundant proteins can be also be used with multiple different techniques. This approach involves breaking the proteome into subproteomes – a group of proteins linked by either cellular location (organelle or specific protein complex, e.g. proteins that bind a specific kinase), or by sharing a chemical propriety (binding ATP or having isoelectric point (pI) values above 9.0). Stasyk & Huber (2004) provide a recent review on fractionation strategies. In terms of broad based analysis, multiple subproteomes may need to be analysed and then the global cellular picture can be put together as the information is integrated.
, 百拇医药
    Qualitative versus quantitative. The next most important choice is whether direct quantification (protein concentration or ratio of protein with respect to one or more condition) of changes is required versus simply analysing the changes qualitatively (all or none or prominent), and whether a goal is to identify post-translational modifications or protein isoform differences. Table 1 illustrates which technologies are appropriate for these choices; however, each has its limitations (see the table and its footnotes). For instance, two-dimensional electrophoresis (2DE) is an excellent choice for relative quantification (by two dimensional difference gel electrophoresis (DIGE) or image analysis software using various dyes), and separation of PTMs (for a review of 2DE, see Gorg et al. 2004); however, the software required for analysis, and scanner equipment needed to take advantage of the robust fluorescent dyes available are very expensive. The length of time required for data analysis and proper spot identification can also be extensive. Liquid chromatography combining chromatographic focusing and hydrophobicity (reversed phase) increases the number of PTMs observed in intact proteins, which can be lost by other methods. It must be recognized that chasing down a PTM, determining the type (phosphorylation versus oxidation or glycosylation) and the exact amino acid residues that are modified, is difficult and often requires additional validation using traditional biochemical methods. Therefore, PTMs are often overlooked and only differences in protein quantification are reported. This is a huge loss to the field; for instance, a method like isotope coded affinity tags (ICAT) with multidimensional liquid chromatography (LC) separation may identify changes in relative abundance of a protein, but may overlook a post-translational modification that is critical in relation to the disease or treatment condition. The future development of more robust methods to track down and monitor PTM is therefore critical. However, the emphasis on developing and expanding the capabilities of each protein separation technology is ongoing and the future is bright. For example see Cantin & Yates (2004) for a recent review on shotgun approaches to PTM analysis.
, 百拇医药
    Relative quantification. Certainly, one of the major breakthroughs in proteomics came with the development of reagents that allow for differential labelling of multiple samples, and then running the samples together in the separation phase (multiplexing). To date, the most successful and widely used examples of this technology is DIGE during 2DE and ICAT using combined chromatography methods. ICAT, which uses different molecular weight mass tags, and affinity purification to attempt to quantify the relative abundance of peptide fragments were the only option available for high-throughput chromatography based applications (Yan et al. 2004; Sethuraman et al. 2004). However, the recent development of iTRAQ, which uses much smaller mass tags of the same molecular weight, but with different fragmentation patterns during tandem MS identification (MS/MS analysis), to determine relative abundance is now another option where the same peptide from different samples is quantified at the same scan in the MS (a limitation of ICAT). Newer reagents are also available including bromine-based mass tags that promise wider applications and ease of use. Table 1 illustrates some of the currently available reagents and their potential applications (for review see Leitner & Lindner 2004).
, 百拇医药
    Mass spectrometry/informatics. Another factor to consider is the degree of difficulty not only in preparation of the sample for separation, but also in the preparation of the fractionated sample for subsequent identification by MS. This can range from simple with methods like reversed phase–HPLC analysis (separating proteins/peptides based on hydrophobicity), since fractions can be easily lyophilized and digested and spotted for either matrix associated laser desorption ionization (MALDI)–time of flight (TOF) or for electrospray ionization (ESI)–MS, to difficult with methods like 2DE, which involves spot cutting, digestion, extraction of peptides and preparation for spotting or MS/MS analysis. Table 1 provides a simplistic summary of the relative difficulty of each method (see Kussmann & Roepstorff (2000) for a review of sample preparation strategies for MALDI).
, http://www.100md.com
    The application of bioinformatics to a proteomics study is unavoidable. The ominous grey box in Fig. 1 illustrates that bioinformatics must be an integrated portion of any proteomics project and will bear different weights at different stages, depending on which separation strategy is utilized. For instance, methods like ICAT require the tracking of information at all stages, and deconvolution of data to reconstruct which peptides match to which proteins, whereas 2DE may require substantial data analysis for absolute and relative quantification. Regardless, all proteomics project will eventually require substantial data management and organization when they reach the protein identification stage. Before starting a proteomics project, one must consider the type of data generated, and how it will be stored, or the risk of becoming overwhelmed with data organization becomes a very real possibility.
, 百拇医药
    MS dominates protein identification, even though immuno-based identification can be used to answer specific questions or be used for validation of ambiguous MS-based identifications. MS identification, regardless of instrumentation can be divided into two strategies, PMF in which purified proteins are identified by the masses of peptide fragments after digestion with different enzymes, or MS/MS analysis. MS/MS utilizes data obtained by instruments with the ability to isolate one peptide and then induce its fragmentation through either its collision with an inert gas (traditional MS/MS analysis), or by molecular energy transfer during laser desorption with MALDI (known as post-source decay or PSD). Immuno-based identification utilizes traditional blotting techniques to confirm MS identifications that are of low statistical confidence. Immuno-chip-based technologies can by-pass the entire need for separation strategies and MS analysis, for the analysis of a known subset of proteins, or as a validation strategy for proteomic data. Briefly, these techniques generally require the adsorption of an antibody (or bait) to a solid phase that is then incubated with a sample and developed similar to an ELISA-based assay (see Table 2 and Espina et al. (2004) for review). Arrays can be used to screen large number of proteins, and holds promise for even broader analysis as number of proteins capable of being observed quantitatively increase into the hundreds.
, 百拇医药
    MS analysis is often the stage at which most proteomics projects fall apart – when a great deal of effort has been applied to separate and isolate proteins, and there is a failure to obtain good MS data. Good quality MS data are required to provide unambiguous protein identification and, if possible, information about modifications to the protein. Unfortunately, MS can be very unforgiving with regard to sample preparation requirements (keratin contamination, protein abundance issues, etc.). Also, as mentioned previously, the question of species (and hence matching to the existing protein/genome databases), separation method (amount of protein or peptides present will impact on the resolution of the separation), and sample preparation method will drive the choice of the type of instrumentation that can be used. Table 3 outlines the ionization sources and detectors available in current mass spectrometers, and some of their strengths and weaknesses. The choice of MS instrumentation available will often influence the project design; for instance, if one is interested in obtaining PTM information on a sequence, the ability to achieve good MS/MS data is essential, although use of this type of instrument generally limits the sample throughput that can be achieved. The MS that most beginners will be exposed to is MALDI-TOF. MALDI-TOF instruments are very user friendly and robust, and result in very accurate masses that can then be used for identification by PMF. However, MALDI is typically limited to complex mixtures of only a few proteins, whereas liquid chromatography (LC) front ends, due to the additional separation capabilities, can often handle even more complex mixtures. Newer instruments combine the benefits of MALDI with the power of obtaining good MS/MS data by using more sophisticated mass analysers.
, http://www.100md.com
    A large degree of effort is expended at the last stage of a proteomics project where identifications must be made from the peak lists obtained from a mass spectrometer. Fenyo (2000) provides a detailed overview of the most common search tools available. Briefly, peak lists are generated by carefully analysing the data generated from a mass spectrometer. For PMF experiments, complex algorithms are used to determine if a peak is real versus electronic noise (peak width, threshold over background, relative ratio of isotopes, mass differences between peaks, etc. are used to make this determination), and then a list is generated for searching. Generally the degree of success on getting a match from the database depends on several factors, including the coverage of the protein digested (actual/theoretical number of fragments), the mass accuracy of the measurement, and the purity of the sample (homogeneity). Recently, guidelines for publishing protein identification data were extablished (Carr et al. 2004). Briefly these include the degree of coverage (number of fragments matched versus theoretical peptides for PMF, and number of fragmentation ions obtained by MS/MS analysis matching the theoretical sequence), the MS analysis software settings used to select peaks for searching, and the parameters used for searching the peak lists against databases.
, 百拇医药
    There is still much advancement required in this area – for example algorithms must be generated to help identify peptide fragments that are modified, databases that include information on small nucleotide polymorphisms, and MS/MS databases that can robustly identify masses corresponding to complex PTMs.

    Collaboration or going it alone

    The last consideration a researcher new to proteomics should consider relates to whether the investigator should choose a collaborator or go it alone with or without the use of a core facility. These choices are typically situation specific, but some general guidelines to consider are the degree of support offered by a core facility, and whether the staff allow participation in a project, or if they will just accept a sample and run it. Often core facilities do not have the time to invest in a substantial degree of optimization of an experimental system, or to necessarily invest the time involved in the extensive data analysis required downstream (at a reasonable cost). This is detrimental to the field of proteomics. Although less common than core facilities, laboratories specializing in proteomics are often a good choice at minimum in an advisory capacity, and often produce very fruitful collaborations. Regardless, the questions for both scenarios are exactly those outlined above, and each stage of the project should be discussed in depth so that reasonable expectations can be realized. Proteomics often requires the juggling of several technologies, so some facilities and proteomics labs may be more experienced at one than another. Additionally, classical biochemistry techniques can be used effectively to complement proteomic discoveries. Often a protein chemistry laboratory that specializes in separation techniques is an ideal initial collaborator, and then the choice remains whether to collaborate or go it alone with protein identification. The last factor is cost and availability of MS instrumentation. MALDI instruments of varying quality are widely available, and more MS/MS instruments are being brought into core facilities, although the sample throughput capabilities of these instruments are much lower. If a project requires an instrument like a Fourier transform ion cyclotron resonance mass spectrometer (FT-ICR MS), there are few groups in the world in academic settings that have these instruments so time on such instruments is limited.
, 百拇医药
    The future

    It is fun, although perhaps dangerous, to speculate on the future in this rapidly changing field, but ideally the future will involve a much higher degree of integration of different experimental approaches, so that minimal experimental findings on a protein's properties (pI, hydrophobicity or molecular mass), plus very limited MS information (one unique fragment) may be used to identify proteins with ease. This will require the development of real-time informatics, to choose the targets that are of high value for mass spectrometry. Another required breakthrough which will assist proteomic experiments is the development of techniques similar to cluster analysis in genomics to determine which classes of proteins are related, so that relative changes between control groups and experimental groups can be differentiated with more ease. Also, as mentioned above, advancements also need to be made with databases to include information on SNPs and PTMs. Finally, as technologies improve to increase the depth of proteome coverage (including instrumentation that can obtain 100% coverage of proteins and interaction mapping projects are completed) interactions between proteins will become more apparent, so that changes in one or two proteins that are visible by current techniques may illuminate changes in other interacting partners that can be chased down using traditional biochemical techniques.
, 百拇医药
    Summary

    Proteomics at its best is the integration of carefully planned research with the advantages of powerful discovery and analytical technologies. If planned properly, proteomic projects can be extremely fruitful, and lay the groundwork for years of further research and investigation. Although it is not high-throughput with respect to the amount of time invested, it certainly is with regard to the breadth of what can be observed. Used appropriately in the context of physiological models of diseases and protein function studies using knockout, knock-in and transgenic expression in animals, several new therapeutic targets and biomarkers can be discovered, and the potential contribution towards the understanding of disease and disease processes is unparalleled. That is the eventual promise of proteomics; the technology is the tool to carry us toward this goal, but it has to be the science that drives the process.
, http://www.100md.com
    References

    Carr S, Aebersold R, Baldwin M, Burlingame A, Clauser K & Nesvizhskii A (2004). The need for guidelines in publication of peptide and protein identification data. Working Group on Publication Guidelines for Peptide and Protein Identification Data. Mol Cell Proteomics, vol. 3, pp. 531–533. Epub 2004 April. 09.

    Cantin GT & Yates JR 3rd (2004). Strategies for shotgun identification of post-translational modifications by mass spectrometry. J Chromatogr A 1053, 7–14.
, http://www.100md.com
    Espina V, Woodhouse EC, Wulfkuhle J, Asmussen HD, Petricoin EF & Liotta LA 3rd (2004). Protein microarray detection strategies: focus on direct detection technologies. J Immunol Meth 290, 121–133.

    Fenyo D (2000). Identifying the proteome: software tools. Curr Opin Biotechnol 11, 391–395.

    Ge Y, Rajkumar L, Guzman RC, Nandi S, Patton WF & Agnew BJ (2004). Multiplexed fluorescence detection of phosphorylation, glycosylation, and total protein in the proteomic analysis of breast cancer refractoriness. Proteomics 4, 3464–3467.
, http://www.100md.com
    Gorg A, Weiss W & Dunn MJ (2004). Current two-dimensional electrophoresis technology for proteomics. Proteomics 15, 15.

    Kussmann M & Roepstorff P (2000). Sample preparation techniques for peptides and proteins analyzed by MALDI-MS. Meth Mol Biol 146, 405–424.

    Leitner A & Lindner W (2004). Current chemical tagging strategies for proteome analysis by mass spectrometry. J Chromatogr B Anal Technol Biomed Life Sci 813, 1–26.
, http://www.100md.com
    Neverova I & Van Eyk JE (2002). Application of reversed phase high performance liquid chromatography for subproteomic analysis of cardiac muscle. Proteomics 2, 22–31.

    O'Farrell PH (1975). High resolution two-dimensional electrophoresis of proteins. J Biol Chem 250, 4007–4021.

    Peng J, Schwartz D, Elias JE, Thoreen CC, Cheng D, Marsischky G, Roelofs J, Finley D & Gygi SP (2003). A proteomics approach to understanding protein ubiquitination. Nat Biotechnol 21, 921–926.
, 百拇医药
    Sethuraman M, McComb ME, Heibeck T, Costello CE & Cohen RA (2004). Isotope-coded affinity tag approach to identify and quantify oxidant-sensitive protein thiols. Mol Cell Proteomics 3, 273–278.

    Shevchenko A, Wilm M, Vorm O & Mann M (1996). Mass spectrometric sequencing of proteins silver-stained polyacrylamide gels. Anal Chem 68, 850–858.

    Stasyk T & Huber LA (2004). Zooming in: Fractionation strategies in proteomics. Proteomics 12, 3704–3716.
, http://www.100md.com
    Steen H & Mann M (2004). The ABC's (and XYZ's) of peptide sequencing. Nat Rev Mol Cell Biol 5, 699–711.

    Wilkins MR & Williams KL (1997). Cross-species protein identification using amino acid composition, peptide mass fingerprinting, isoelectric point and molecular mass: a theoretical evaluation. J Theor Biol 186, 7–15.

    Yan W, Lee H, Deutsch EW, Lazaro CA, Tang W, Chen E, Fausto N, Katze MG & Aebersold R (2004). A dataset of human liver proteins identified by protein profiling via isotope-coded affinity tag (ICAT) and tandem mass spectrometry. Mol Cell Proteomics 3, 1039–1041., 百拇医药(David R. M Graham,, Steve)