Big Chemical Encyclopedia

Chemical substances, components, reactions, process design ...

Articles Figures Tables About

Principal component analysis scaling

One of the main attractions of normal mode analysis is that the results are easily visualized. One can sort the modes in tenns of their contributions to the total MSF and concentrate on only those with the largest contributions. Each individual mode can be visualized as a collective motion that is certainly easier to interpret than the welter of information generated by a molecular dynamics trajectory. Figure 4 shows the first two normal modes of human lysozyme analyzed for their dynamic domains and hinge axes, showing how clean the results can sometimes be. However, recent analytical tools for molecular dynamics trajectories, such as the principal component analysis or essential dynamics method [25,62-64], promise also to provide equally clean, and perhaps more realistic, visualizations. That said, molecular dynamics is also limited in that many of the functional motions in biological molecules occur in time scales well beyond what is currently possible to simulate. [Pg.165]

Because of the relatively small number of experiments done on commercial-scale equipment before submission, and the often very narrow factor ranges (Hi/Lo might differ by only 5-10%), if conditions are not truly under control, high-level models (multi-variate regressions, principal components analysis, etc.) will pick up spurious signals due to noise and unrecognized drift. For example, Fig. 4.43 summarizes the yields achieved for... [Pg.303]

Univac. Large scale systems, STAT-PACK, FACTAN - Factor and principal component analysis (1973) 33-39. [Pg.939]

Computational methods have been applied to determine the connections in systems that are not well-defined by canonical pathways. This is either done by semi-automated and/or curated literature causal modeling [1] or by statistical methods based on large-scale data from expression or proteomic studies (a mostly theoretical approach is given by reference [2] and a more applied approach is in reference [3]). Many methods, including clustering, Bayesian analysis and principal component analysis have been used to find relationships and "fingerprints" in gene expression data [4]. [Pg.394]

The difference between interval and ratio scales can be important for including or not including an intercept term in mathematical models for the correct calculation of the correlation coefficient for deciding to mean center or not in principal component analysis and for a host of other decisions in data treatment and modeling. [Pg.19]

In some diseases a simple ordinal scale or a VAS scale cannot describe the full spectrum of the disease. There are many examples of this including depression and erectile dysfunction. Measurement in such circumstances involves the use of multiple ordinal rating scales, often termed items. A patient is scored on each item and the summation of the scores on the individual items represents an overall assessment of the severity of the patient s disease status at the time of measurement. Considerable amoimts of work have to be done to ensure the vahdity of these complex scales, including investigations of their reprodu-cibihty and sensitivity to measuring treatment effects. It may also be important in international trials to assess to what extent there is cross-cultural imiformity in the use and imderstand-ing of the scales. Complex statistical techniques such as principal components analysis and factor analysis are used as part of this process and one of the issues that need to be addressed is whether the individual items should be given equal weighting. [Pg.280]

Principal component analysis (PCA) is another tool that has been used extensively to analyze molecular simulations. The technique, which attempts to extract the large-scale characteristic motions from a structural ensemble, was first applied to biomolecular simulations by Garcia [28], although an analogous technique was used by Levy et al. [30]. The first step is the construction of the 3N x 3N (for an N-atom system) fluctuation correlation matrix... [Pg.39]

How is dimension reduction of chemical spaces achieved There are a number of different concepts and mathematical procedures to reduce the dimensionality of descriptor spaces with respect to a molecular dataset under investigation. These techniques include, for example, linear mapping, multidimensional scaling, factor analysis, or principal component analysis (PCA), as reviewed in ref. 8. Essentially, these techniques either try to identify those descriptors among the initially chosen ones that are most important to capture the chemical information encoded in a molecular dataset or, alternatively, attempt to construct new variables from original descriptor contributions. A representative example will be discussed below in more detail. [Pg.282]

Because principal component analysis attempts to account for all of the variance within a molecular dataset, it can be negatively affected by outliers, i.e., compounds having at least some descriptor values that are very different from others. Therefore, it is advisable to scale principal component axes or, alternatively, pre-process compound collections using statistical filters to identify and remove such outliers prior to the calculation of principal components. [Pg.287]

Some of the linear combinations will be well defined and others poorly defined. The latter may be eliminated in a filtering procedure, referred to in the literature under the names characteristic value filtering, eigenvalue filtering, and principal component analysis. If the parameter set is not homogeneous, but includes different types, relative scaling is important. Watkin (1994) recommends that the unit be scaled such that similar shifts in all parameters lead to similar changes in the error function S. [Pg.79]

Spectral data are highly redundant (many vibrational modes of the same molecules) and sparse (large spectral segments with no informative features). Hence, before a full-scale chemometric treatment of the data is undertaken, it is very instructive to understand the structure and variance in recorded spectra. Hence, eigenvector-based analyses of spectra are common and a primary technique is principal components analysis (PC A). PC A is a linear transformation of the data into a new coordinate system (axes) such that the largest variance lies on the first axis and decreases thereafter for each successive axis. PCA can also be considered to be a view of the data set with an aim to explain all deviations from an average spectral property. Data are typically mean centered prior to the transformation and the mean spectrum is used a base comparator. The transformation to a new coordinate set is performed via matrix multiplication as... [Pg.187]

Relationships between the individual LOE can be examined via principal components analysis (PCA). Correlations among principal components for individual LOE indicate concordance or agreement. Relationships between different SQT LOE can also be assessed using other methods including Mantel s test (Legendre and Fortin, 1989) coupled with a measure of similarity or ordination canonical discriminant (or correspondence) analyses multidimensional scaling (MDS). [Pg.313]

Relationship between Palatability and Umami. Yamanaka et at. (40) collected words expressing "palatability". They did this by asking people to write down their definition of palatability, excluding appearance, aroma and texture. From the total of 1900 expressions obtained, 38 of them were selected as important. The similality between each pair of the expressions was measured on a 5-point scale using a mass panel. The data obtained were analyzed by principal component analysis and cluster analysis. As a result, concrete expressions of palatability were classified into the following five groups ... [Pg.47]

Comparison and ranking of sites according to chemical composition or toxicity is done by multivariate nonparametric or parametric statistical methods however, only descriptive methods, such as multidimensional scaling (MDS), principal component analysis (PCA), and factor analysis (FA), show similarities and distances between different sites. Toxicity can be evaluated by testing the environmental sample (as an undefined complex mixture) against a reference sample and analyzing by inference statistics, for example, t-test or analysis of variance (ANOVA). [Pg.145]

Chemists and statisticians use the term mixture in different ways. To a chemist, any combination of several substances is a mixture. In more formal statistical terms, however, a mixture involves a series of factors whose total is a constant sum this property is often called closure and will be discussed in completely different contexts in the area of scaling data prior to principal components analysis (Chapter 4, Section 4.3.6.5 and Chapter 6, Section 6.2.3.1). Hence in statistics (and chemometrics) a solvent system in HPLC or a blend of components in products such as paints, drugs or food is considered a mixture, as each component can be expressed as a proportion and the total adds up to 1 or 100%. The response could be a chromatographic separation, the taste of a foodstuff or physical properties of a manufactured material. Often the aim of experimentation is to find an optimum blend of components that tastes best, or provide die best chromatographic separation, or die material diat is most durable. [Pg.84]

Z-scales are obtained by principal component analysis of physico-chemical properties of monomers. E.g., the first Z-scales of amino adds describe hydrophobicity (z1), steric bulk/polarisability (z2) and polarity (z3) of the amino acids. [Pg.293]

Preference mapping can be accomplished with projection techniques such as multidimensional scaling and cluster analysis, but the following discussion focuses on principal components analysis (PCA) [69] because of the interpretability of the results. A PCA represents a multivariate data table, e.g., N rows ( molecules ) and K columns ( properties ), as a projection onto a low-dimensional table so that the original information is condensed into usually 2-5 dimensions. The principal components scores are calculated by forming linear combinations of the original variables (i.e., properties ). These are the coordinates of the objects ( molecules ) in the new low-dimensional model plane (or hyperplane) and reveal groups of similar... [Pg.332]


See other pages where Principal component analysis scaling is mentioned: [Pg.697]    [Pg.421]    [Pg.446]    [Pg.477]    [Pg.123]    [Pg.41]    [Pg.67]    [Pg.396]    [Pg.185]    [Pg.20]    [Pg.401]    [Pg.85]    [Pg.107]    [Pg.109]    [Pg.114]    [Pg.3]    [Pg.144]    [Pg.37]    [Pg.618]    [Pg.171]    [Pg.305]    [Pg.10]    [Pg.7]    [Pg.356]    [Pg.157]    [Pg.217]    [Pg.190]    [Pg.3]    [Pg.47]    [Pg.292]    [Pg.136]   
See also in sourсe #XX -- [ Pg.111 , Pg.112 ]




SEARCH



Component analysis

Principal Component Analysis

Principal analysis

Principal component analysi

Scaling analysis

© 2024 chempedia.info