Big Chemical Encyclopedia

Chemical substances, components, reactions, process design ...

Articles Figures Tables About

Data correlation matrix

Finally, Schlieren effect minimisation is inherent to most strategies already proposed for multivariate analyses relying on spectral exploitation. In fact, the calibration models are usually built after data decorrelation. As the Schlieren noise superimposed on every data point leads to an enhancement of the data correlation matrix, procedures for data decorrelation implicitly lead to Schlieren minimisation. This aspect has been demonstrated in relation to the partial least squares algorithm [120]. [Pg.137]

In principle, this set of equations can be solved for the various constants, a through Q, just as a and b were obtained previously. In practice, however, the actual numerical evaluation involves considerable computation in all but the simplest examples. Computer solution by matrix techniques designed specifically to handle this type of data correlation problem is usually required. [Pg.245]

Kelkar and McCarthy (1995) proposed another method to use the feedforward experiments to develop a kinetic model in a CSTR. An initial experimental design is augmented in a stepwise manner with additional experiments until a satisfactory model is developed. For augmenting data, experiments are selected in a way to increase the determinant of the correlation matrix. The method is demonstrated on kinetic model development for the aldol condensation of acetone over a mixed oxide catalyst. [Pg.143]

The correlation coefficients can be arranged in a matrix like the covariances. The resulting correlation matrix (R, with l s in the main diagonal) is for autoscaled x-data identical to C. [Pg.56]

Based on the correlation matrix of all bioassays data obtained with 37 effluents, it can be concluded that none of the bioassays produces data that are redundant. In other words, all bioassay procedures add to the information content of the PEEP index. [Pg.42]

If more than 2 variables are involved, the correlation matrix must be positive, semidefinite. That means if the correlation with A and Bis a, and the correlation between B and C is b, then the correlation between A and C can t be any number between 0 and 1 it must satisfy certain constraints. Such errors may occur if pairwise correlation coefficients stem from different data sets. [Pg.158]

In total, 4,777 solutes and 10,198 IgD values were analyzed. From these data, the correlation matrix of size 49 X 49 (48 conventional solvents + IL) was derived. It appeared that the solvents most close to [C4CiIm][PFJ (i.e., having the highest pairwise correlation coefficient solvent/IL) are esters with a short alkyl chain (ethyl acetate correlation coefficient r = 0.95, as determined by the distribution ratios for n = 11 solutes butyl acetate r = 0.92, n = 30) and substituted aromatic hydrocarbons (m-xylene r = 0.92, n = 20). The most distant from ILs are aliphatic hydrocarbons. Interestingly, the correlation with 1-octanol is moderate, r = 0.76, n = 56. [Pg.252]

The decomposition according to Eq. 5-16 is performed by a principal axes transformation of the correlation matrix R. The correlation matrix of the raw data is therefore the starting point of the calculations. [Pg.165]

The reader has seen in Section 5.4.1 that the total number of eigenvalues and eigenvectors, which is the same as the number of features, reproduces the correlation matrix and therefore describes the total variance of the data. The general model of both PCA and FA described in Section 5.4.1 is therefore called complete factor solution. [Pg.171]

The regression analysis of multicollinear data is described in several papers e.g. [MANDEL, 1985 HWANG and WINEFORDNER, 1988], HWANG and WINEFORD-NER [1988] also discuss the principle of ridge regression, which is, essentially, the addition of a small contribution to the diagonal of correlation matrix. The method of partial least squares (PLS) described in Section 5.7.2 is one approach to solving this problem. [Pg.197]

Preliminary data analysis carried out for the spectral datasets were functional group mapping, and/or hierarchical cluster analysis (HCA). This latter method, which is well described in the literature,4,9 is an unsupervised approach that does not require any reference datasets. Like most of the multivariate methods, HCA is based on the correlation matrix Cut for all spectra in the dataset. This matrix, defined by Equation (9.1),... [Pg.193]

Note that a PCA analysis often starts by prestandardizing the data to obtain variables that all have the same spread. Otherwise, the variables with a large variance compared with the others will dominate the first principal components. Standardizing by the mean and the standard deviation of each variable yields a PCA analysis based on the correlation matrix instead of the covariance matrix. We can also standardize each variable j in a robust way, e.g., by first subtracting its median, medQty,. .., x,v), and then dividing by its robust scale estimate, Qn(Xip > -%) ... [Pg.189]

The analyst should check the Shepard diagram that represents a step line so-called D-hat values. If all reproduced distances fall onto the step-line, then the rank ordering of distances (or similarities) would be perfectly reproduced by the dimensional model, while deviations from the step-line mean lack of fit. The interpretation of the dimensions usually represents the final step of this multivariate procedure. As in factor analysis, the final orientation of axes in the plane (or space) is mostly the result of a subjective decision by the researcher since the distances between objects remain invariable regardless of the type of the rotation. However, it must be remembered that MDS and FA are different methods. FA requires that the underlying data be distributed as multivariate normal, whereas MDS does not impose such a restriction. MDS often yields more interpretable solutions than FA because the latter tends to extract more factors. MDS can be applied to any kind of distances or similarities (those described in cluster analysis), whereas FA requires firstly the computation of the correlation matrix. Figure 7.3 shows the results of applying MDS to the samples described in the CA and FA sections (7.3.1 and 7.3.2). [Pg.165]

The variances and covariance of the descriptors are given by the matrix (X — X) (X — X), in which the diagonal elements are the variances of the variables and the off-diagonal elements are the covariances. When the data have been scaled to unit variance, this matrix is called the correlation matrix and the off-diagonal elements are correlation coefficients for the correlations between the variables, and the sum of the variances is equal to the number of variables. [Pg.37]

Table 3 Cross-Correlation Matrix for Neutral Solute Data in Table... [Pg.37]

Table 4 Cross-Correlation Matrix for Ionic Solute Data in Table 2 ... Table 4 Cross-Correlation Matrix for Ionic Solute Data in Table 2 ...

See other pages where Data correlation matrix is mentioned: [Pg.419]    [Pg.420]    [Pg.413]    [Pg.83]    [Pg.47]    [Pg.137]    [Pg.320]    [Pg.256]    [Pg.20]    [Pg.6]    [Pg.250]    [Pg.236]    [Pg.267]    [Pg.53]    [Pg.160]    [Pg.62]    [Pg.269]    [Pg.530]    [Pg.282]    [Pg.17]    [Pg.58]    [Pg.29]    [Pg.166]    [Pg.274]    [Pg.217]    [Pg.163]    [Pg.253]    [Pg.92]    [Pg.164]    [Pg.62]    [Pg.68]    [Pg.142]    [Pg.122]   
See also in sourсe #XX -- [ Pg.140 ]




SEARCH



Correlation matrix

Correlative data

Data matrix

© 2024 chempedia.info