Big Chemical Encyclopedia

Chemical substances, components, reactions, process design ...

Articles Figures Tables About

Computational training data sets

The rest of the paper is organized as follows. The Section 2 describes attack classification and training data set. In the Section 3 the intrusion detection system is described, based on neural network approach. Section 4 presents the nonlinear PCA neural network and multilayer perceptron for identification and classification of computer network attack. In Section 5 the results of experiments are presented. Conclusion is given in Section 6. [Pg.368]

The other method is a computer- or statistically-based expert system, for which a large training data set of compounds with known toxicity is needed to derive structural features that are highly correlated to the specified toxicity (Durham and Pearl 2001). An example of a typical computer- or statistically-based expert system is MultiCASE, which will also be described later. [Pg.801]

Validation without an independent test set. Each application of the adaptive wavelet algorithm has been applied to a training set and validated using an independent test set. If there are too few observations to allow for an independent testing and training data set, then cross validation could be used to assess the prediction performance of the statistical method. Should this be the situation, it is necessary to mention that it would be an extremely computational exercise to implement a full cross-validation routine for the AWA. That is. it would be too time consuming to leave out one observation, build the AWA model, predict the deleted observation, and then repeat this leave-one-out procedure separately. In the absence of an independent test set, a more realistic approach would be to perform cross-validation using the wavelet produced at termination of the AWA, but it is important to mention that this would not be a full validation. [Pg.200]

Predictive methods aimed at extracting secondary structural information from the linear primary sequence make extensive use of neural networks, traditionally used for analysis of patterns and trends. Basically, a neural network provides computational processes the ability to learn in an attempt to approximate human learning versus following instructions blindly in a sequential marmer. Every nemal network has an input layer and an output layer. In the case of secondary structure prediction, the input layer would be information from the sequence itself, and the output layer would be the probabilities of whether a particular residue could form a particular structure. Between the input and output layers would be one or more hidden layers where the actual learning would take place. This is accomplished by providing a training data set for the network. Here, an appropriate training set would be all sequences for which three-dimensional structures have been deduced. The network can process this information to look for what are possibly weak relationships between an amino acid sequence and the structures they can form in a particular context. A more complete discussion of neural networks as applied to secondary structure prediction can be found in Kneller et al. (1990). [Pg.264]

The framework of presented intelligent multi-sensor system is reflected by its data processing flow as illustrated in Fig. 3. Diversified sensors in field and sophisticated algorithms make the system scalable and adaptive to different driving profiles and scenarios. Data sets of complementary sensors are synchronized on the same time base before being conveyed to feature computation components. Based on the outcome of feature computation selected data sets are fused on the Mature level to construct input vectors for pattern classification so as to detect driver drowsiness. The classifier being used in this work is built upon Artificial Neural Network (ANN) or, more particularly. Multilayer Perceptrons (MLP) with supervised training procedure. [Pg.126]

In Example 4.9, the results from Example 4.8 are used to compute the residual variance and / -ratios for the data set described in Figure 4.16. The / -values for the 10 unknown spectra are shown in Table 4.2. The unknown spectrum contaminated with a minor level of an impurity is shown in the first row. All samples in the training set have small residual variances and / -ratios less than the critical value of F = 4.105. The unacceptable unknown spectrum has a very large F-value, indicating with a high degree of confidence that it is not a member of the parent population represented by the training set. [Pg.101]


See other pages where Computational training data sets is mentioned: [Pg.26]    [Pg.229]    [Pg.388]    [Pg.235]    [Pg.455]    [Pg.574]    [Pg.759]    [Pg.59]    [Pg.312]    [Pg.679]    [Pg.692]    [Pg.168]    [Pg.213]    [Pg.1094]    [Pg.14]    [Pg.310]    [Pg.161]    [Pg.201]    [Pg.80]    [Pg.575]    [Pg.5]    [Pg.337]    [Pg.140]    [Pg.454]    [Pg.462]    [Pg.159]    [Pg.466]    [Pg.518]    [Pg.524]    [Pg.532]    [Pg.544]    [Pg.161]    [Pg.97]    [Pg.398]    [Pg.31]    [Pg.143]    [Pg.191]    [Pg.317]    [Pg.367]    [Pg.176]    [Pg.247]    [Pg.293]    [Pg.64]    [Pg.346]    [Pg.681]   
See also in sourсe #XX -- [ Pg.26 , Pg.27 ]




SEARCH



Data set

Training data

Training set

© 2024 chempedia.info