Big Chemical Encyclopedia

Chemical substances, components, reactions, process design ...

Articles Figures Tables About

Normal form error correction

A table is said to be in first normal form if each row has the same number of columns, each column has a value, and there are no duplicate rows. Because an RDBMS uses a table defined with a fixed number of columns, it is always true that each row contains the same number of columns. If one allows that null is a value, then every column will have a value. It should be obvious that repeating a row in a table is wasteful, but also potentially confusing and prone to error. For example, if two rows in a table of logP contained the same name and logP, one row may have the logP changed at some point. Then which row would be the correct row This condition also illustrates the final aspect of first normal form There should be at least one column, or combination of columns, that could function as a key that uniquely identifies the row. This is the name column or compound id column in the above examples. The data in this column must be unique. [Pg.17]

The COMPU-RATE, developed by Faehr Electronic Timers, Inc., is a portable device using batteries, which provide about 120 hr of running time. Manual entries are required only for the element column and the top four lines of the form. This allows the analyst to concentrate on observing the work and the operator performance. Time can be in thousandths of a minute or one hundred-thousandths of an hour. The COMPU-RATE software system computes the mean and median element values, adjustment of mean times to normal time after inputting the performance rating, and allowed times in minutes and/or hours per piece and pieces per hour. Errors can be corrected through an edit function. [Pg.1414]

As a safeguard for cases, where the correct normal form has not been included into the set ACF, one should monitor the approximation error err of the best neural normal form model. [Pg.162]

Furthermore, it is sometimes questionable to use literature data for modeling purposes, as small variations in process parameters, reactor hydrodynamics, and analytical equipment limitations could skew selectivity results. To obtain a full product spectrum from an FT process, a few analyses need to be added together to form a complete picture. This normally involves analysis of the tail gas, water, oil, and wax fractions, which need to be combined in the correct ratio (calculated from the drainings of the respective phases) to construct a true product spectrum. Reducing the number of analyses to completely describe the product spectrum is one obvious way to minimize small errors compounding into large variations in... [Pg.231]

Although monoliths have many attractive features, their synthesis can prove troublesome. It is often a matter of trial and error to find the correct combination of monomer, cross-linking agent, poragen, initiator and polymerization temperature to obtain workable results. This optimization process must be repeated each time until ideal conditions are found [114,115]. However, this process allows for the formation of a monolith of virtually any shape desired, normally some form of secondary containment, such as a column. [Pg.94]

The characters are first normalized by rotating the original scanned image to correct for scanning error and by combinations of scaling under sampling and contrast and density adjustments of the scanned characters. In operation, the normalized characters are then presented to a multilayer perceptron neural network for recognition the network was trained on exemplars of characters form numerous serif and sans serif fonts to achieve font invariance. Where the output from the neural network indicates more than one option, for example 5 and s, the correct interpretation is determined from context. [Pg.56]

In an early attempt to calculate the phase fractions in an approximate implicit volume fraction-velocity-pressure correction procedure, Spalding [176, 177, 178, 180] calculated the phase fractions from the respective phase continuity equations. However, experience did show that it was difficult to conserve mass simultaneously for both phases when the algorithm mentioned above was used. For this reason, Spalding [179] suggested that the volume fraction of the dispersed phase may rather be calculated from a discrete equation that is derived from a combination of the two continuity equations. An alternative form of the latter volume fraction equation, particularly designed for fluids with large density differences, was later proposed by Carver [26]. In this method the continuity equations for each phase were normalized by a reference mass density to balance the weight of the error for each phase. [Pg.1061]

The derivatives of the calculated data with respect to the force constants are then formed, and these are used to construct the normal equations from which corrections to the force constants are calculated in such a way as to minimize the sum of weighted squares of residuals. Because the relations between the data and the force constants are often very nonlinear, it is necessary to cycle this calculation until the changes in the force constants drop to zero when the calculation will have converged and the sum of weighted squares of errors will be minimized. The usual statistical formulas are then used to obtain the variance/covariance matrix in the derived best estimates of the force constants, and the estimated standard errors in the force constants are usually quoted along with their values. The whole procedure is referred to as a force constant refinement calculation. ... [Pg.284]

No information presented in the reference document suggests typical values for errors associated with the concentration measurements, or the type of statistical distribution that might represent these errors. The document does suggest that the preferred measurement of error is a 95-percent confidence limit based on the t-statistic, implying that the error distribution is normal in appearance. The equation presented to compute the 95-percent confidence limit was incorrect in the reference. The following equation is the correct form ... [Pg.58]

Two of the most common systematic errors that occur in experimental total structure factors are small normalization errors in the form of additive and multiplicative constants, particularly for X-ray data because of the Q dependence of the form factor. Such errors can be accounted for within the RMC algorithm, since the RMC structure factors are correctly normalized (given the correct density, atomic composition, etc.). The required multiplicative factor which minimizes x2 is given by... [Pg.154]


See other pages where Normal form error correction is mentioned: [Pg.4522]    [Pg.30]    [Pg.4521]    [Pg.87]    [Pg.162]    [Pg.154]    [Pg.162]    [Pg.242]    [Pg.79]    [Pg.110]    [Pg.4]    [Pg.253]    [Pg.160]    [Pg.724]    [Pg.22]    [Pg.112]    [Pg.530]    [Pg.90]    [Pg.265]    [Pg.442]    [Pg.296]    [Pg.54]    [Pg.77]    [Pg.53]    [Pg.42]    [Pg.176]    [Pg.450]    [Pg.214]    [Pg.190]    [Pg.59]    [Pg.961]    [Pg.283]    [Pg.4]    [Pg.52]    [Pg.404]    [Pg.208]    [Pg.144]    [Pg.190]    [Pg.526]   
See also in sourсe #XX -- [ Pg.176 ]




SEARCH



Errors corrections

Errors normal

Normal form

Normal form correction

© 2024 chempedia.info