Big Chemical Encyclopedia

Chemical substances, components, reactions, process design ...

Articles Figures Tables About

Number of data points

The computer algorithms that carry out the discrete Fourier transform calculation work most efficiently if the number of data points (np) is an integral power of 2. Generally, for basic H and C spectra, at least 16,384 (referred to as 16K ) data points, and 32,768 ( 32K ) points should be collected for full H and spectral windows, respectively. With today s higher field instruments and large-memory computers, data sets of 64K for H and 64-128K for and other nuclei are now commonly used. [Pg.41]


The most reliable estimates of the parameters are obtained from multiple measurements, usually a series of vapor-liquid equilibrium data (T, P, x and y). Because the number of data points exceeds the number of parameters to be estimated, the equilibrium equations are not exactly satisfied for all experimental measurements. Exact agreement between the model and experiment is not achieved due to random and systematic errors in the data and due to inadequacies of the model. The optimum parameters should, therefore, be found by satisfaction of some selected statistical criterion, as discussed in Chapter 6. However, regardless of statistical sophistication, there is no substitute for reliable experimental data. [Pg.44]

This sum, when divided by the number of data points minus the number of degrees of freedom, approximates the overall variance of errors. It is a measure of the overall fit of the equation to the data. Thus, two different models with the same number of adjustable parameters yield different values for this variance when fit to the same data with the same estimated standard errors in the measured variables. Similarly, the same model, fit to different sets of data, yields different values for the overall variance. The differences in these variances are the basis for many standard statistical tests for model and data comparison. Such statistical tests are discussed in detail by Crow et al. (1960) and Brownlee (1965). [Pg.108]

Appendix C-6 gives parameters for all the condensable binary systems we have here investigated literature references are also given for experimental data. Parameters given are for each set of data analyzed they often reflect in temperature (or pressure) range, number of data points, and experimental accuracy. Best calculated results are usually obtained when the parameters are obtained from experimental data at conditions of temperature, pressure, and composition close to those where the calculations are performed. However, sometimes, if the experimental data at these conditions are of low quality, better calculated results may be obtained with parameters obtained from good experimental data measured at other conditions. [Pg.144]

Steps 2 and 3 are performed for all input objects. When all data points have been fed into the network one training epoch has been achieved. A network is usually trained in several training epochs, depending on the size of the network and the number of data points. [Pg.457]

It should require as small a number of data points as possible to achieve an accurate fit. [Pg.176]

The degree of data spread around the mean value may be quantified using the concept of standard deviation. O. If the distribution of data points for a certain parameter has a Gaussian or normal distribution, the probabiUty of normally distributed data that is within Fa of the mean value becomes 0.6826 or 68.26%. There is a 68.26% probabiUty of getting a certain parameter within X F a, where X is the mean value. In other words, the standard deviation, O, represents a distance from the mean value, in both positive and negative directions, so that the number of data points between X — a and X -H <7 is 68.26% of the total data points. Detailed descriptions on the statistical analysis using the Gaussian distribution can be found in standard statistics reference books (11). [Pg.489]

Time constraints ate an important factor in selecting nmr experiments. There are four parameters that affect the amount of instmment time requited for an experiment, A preparation delay of 1—3 times should be used. Too short a delay results in artifacts showing up in the 2-D spectmm whereas too long a delay wastes instmment time. The number of evolution times can be adjusted. This affects the F resolution. The acquisition time or number of data points in can be adjusted. This affects resolution in F. EinaHy, the number of scans per EID can be altered. This determines the SNR for the 2-D matrix. In general, a lower SNR is acceptable for 2-D than for 1-D studies. [Pg.408]

Correlations based on reference substances are limited to compounds which have experimentally determined values, but the number of data points needed to produce a correlation is relatively small. A reference substance should be as chemically and physically compatible to the chemical with which it is being compared as possible. Use of reference substances in the ideal assumes no property deviations, thus the smaller the deviations, the lower the absolute error in the correlations. [Pg.242]

Number of data points, number of stages or effects Number of inputs/outputs, model horizon Pressure... [Pg.717]

The historical data is sampled at user-specified intervals. A typical process plant contains a large number of data points, but it is not feasible to store data for all points at all times. The user determines if a data point should be included in the list of archive points. Most systems provide archive-point menu displays. The operators are able to add or delete data points to the archive point hsts. The samphng periods are normally some multiples of their base scan frequencies. However, some systems allow historical data samphng of arbitraiy intei vals. This is necessaiy when intermediate virtual data points that do not have the scan frequency attribute are involved. The archive point lists are continuously scanned bv the historical database software. On-line databases are polled for data. The times of data retrieval are recorded with the data ootained. To consei ve storage space, different data compression techniques are employed by various manufacturers. [Pg.773]

The parameters of the Dirichlet prior for the q s should be proportional to the counts for each component in this preliminary data analysis. So we now have a collection of prior parameters 6oi = ( J.oi, Kq , Go , Vo ) and a preliminary assignment of each data point to a component, cj, and therefore the preliminary number of data points for each component, A . ... [Pg.328]

The constant of proportionality is based on nonnalizing the probability and establishing the size of the prior, that is, the number of data points that the prior represents. The advantage of the Dirichlet formalism is that it gives values for not only the modes of the probabilities but also the variances, covariances, etc. See Eq. (13). [Pg.339]

Number and type of record The number of data points or tables of data presented in the resource or the number of events the data set reflects where available, the form in which the data are presented, such as failure rates or availability data, confidence intervals or error factors the raw data source used, sueh as surveys, plant records, tests, or judgment. [Pg.29]

When applied to the SEC column, the calibrated polydisperse polymer solution provides a large number of data points in a single run. Use of a standard with a molecular size distribution that encompasses the full separation range for the column allows the entire separation range to be calibrated in a single run (Fig. 2.4). [Pg.33]

Number of data point columns = modulation time x sampling rate... [Pg.93]

Ideally, there should be >4 data points for each estimated parameter. Under this guideline, a 3 parameter logistic function should have 12 data points. At the least, the number of data points minus the number of parameters should be > 3. [Pg.237]

SSqs (clfs) is number of data points minus the common max, common slope, and four fitted values for ECso- Thus, clfs = 24 — 6=18. The value for F for comparison of the simple model (common maximum and slope) to the complex model (individual maxima and slopes) for the data shown in Figure 11.14 is F = 2.4. To be significant at the 95% level of confidence (5% chance that this F actually is not significant), the value of F for df =12, 18 needs to be >2.6. Therefore, since F is less than this value there is no statistical validation for usage of the most complex model. The data should then be fit to a family of curves of common maximum and slope and the individual ECS0 values used to calculate values of DR. [Pg.243]

An alternative method of variance scaling is to scale each variable to a uniform variance that is not equal to unity. Instead we scale each data point by the root mean squared variance of all the variables in the data set This is, perhaps, the most commonly employed type of variance scaling because it is a bit simpler and faster to compute. A data set scaled in this way will have a total variance equal to the number of variables in the data set divided by the number of data points minus one. To use this method of variance scaling, we compute a scale factor, sr, over all of the variables in the data matrix, 8g,... [Pg.177]

Figure 3.12 Comparison of the chromatographic peak shapes obtained from (a) analogue and (b,c) digital detectors, and the effect on peak shape of the number of data points defining the signal. Figure 3.12 Comparison of the chromatographic peak shapes obtained from (a) analogue and (b,c) digital detectors, and the effect on peak shape of the number of data points defining the signal.
Experience has shown that correlations of good precision are those for which SD/RMS. 1, where SD is the root mean square of the deviations and RMS is the root mean square of the data Pfs. SD is a measure equal to, or approaching in the limit, the standard deviation in parameter predetermined statistics, where a large number of data points determine a small number of parameters. In a few series, RMS is so small that even though SD appears acceptable, / values do exceed. 1. Such sets are of little significance pro or con. Evidence has been presented (2p) that this simple / measure of statistical precision is more trustworthy in measuring the precision of structure-reactivity correlations than is the more conventional correlation coefficient. [Pg.16]

Therefore, the velocities of liquid are consistent with the velocities of particles, that is, the motion of nano-particles can reflect the flow of liquid verily in the given condition. Figures 39 and 40 show the comparison of the two liquid samples with different mass concentration of the nanoparticles at different flow rates. Generally, particle velocity increases synchronously with the liquid flow rate, but the velocity becomes dispersive when it exceeds 300 /u,L/min. The more the particles were added in the liquid, the more dispersive of the velocities of the particles were observed. Several possible causes can result in this phenomenon. One possible reason is that when the velocity of flow becomes large enough, the bigger particles in the liquid cannot follow the flow as the smaller particles do, or bigger particles will move slower than the liquid around them, so the velocities of particles will distribute dispersedly. Another possible reason is that when the velocity of flow increases the time for particles to traverse, the view field of the microscope will decrease. As a result, the number of data points in the trace of a particle... [Pg.27]

Because the number of data points is low, many of the statistical techniques that are today being discussed in the literature caimot be used. While this is true for the vast majority of control work that is being done in industrial labs, where acceptability and ruggedness of an evaluation scheme are major concerns, this need not be so in R D situations or exploratory or optimization work, where statisticians could well be involved. For products going to clinical trials or the market, the liability question automatically enforces the tried-and-true sort of solution that can at least be made palatable to lawyers on account of the reams of precedents, even if they do not understand the math involved. [Pg.11]

Figure 1.8. Schematic frequency distributions for some independent (reaction input or control) resp. dependent (reaction output) variables to show how non-Gaussian distributions can obtain for a large population of reactions (i.e., all batches of one product in 5 years), while approximate normal distributions are found for repeat measurements on one single batch. For example, the gray areas correspond to the process parameters for a given run, while the histograms give the distribution of repeat determinations on one (several) sample(s) from this run. Because of the huge costs associated with individual production batches, the number of data points measured under closely controlled conditions, i.e., validation runs, is miniscule. Distributions must be estimated from historical data, which typically suffers from ever-changing parameter combinations, such as reagent batches, operators, impurity profiles, etc. Figure 1.8. Schematic frequency distributions for some independent (reaction input or control) resp. dependent (reaction output) variables to show how non-Gaussian distributions can obtain for a large population of reactions (i.e., all batches of one product in 5 years), while approximate normal distributions are found for repeat measurements on one single batch. For example, the gray areas correspond to the process parameters for a given run, while the histograms give the distribution of repeat determinations on one (several) sample(s) from this run. Because of the huge costs associated with individual production batches, the number of data points measured under closely controlled conditions, i.e., validation runs, is miniscule. Distributions must be estimated from historical data, which typically suffers from ever-changing parameter combinations, such as reagent batches, operators, impurity profiles, etc.
N is the number of data points M is the number of peaks Peakloc(j) is the location of the peak Intensity(j) = lntegral(j)/(pi linewidth)... [Pg.164]


See other pages where Number of data points is mentioned: [Pg.240]    [Pg.44]    [Pg.715]    [Pg.694]    [Pg.59]    [Pg.400]    [Pg.400]    [Pg.408]    [Pg.408]    [Pg.244]    [Pg.459]    [Pg.321]    [Pg.9]    [Pg.241]    [Pg.241]    [Pg.144]    [Pg.70]    [Pg.80]    [Pg.212]    [Pg.826]    [Pg.167]    [Pg.317]    [Pg.387]    [Pg.35]    [Pg.42]   


SEARCH



Data points

Discrepancies or insufficient number of data points

Number of points

© 2024 chempedia.info