Search:

  • Data Reduction

Search Options

Sort by:

Publication Type:

Open access:

Publication Date:

Periodicals:

Search results

Online since: December 2013
Authors: Zhi Bo Ren, Lei Sun, Chun Miao Yan, Yu Zhou Wei
There are a lot of continuous attributes in data streams, and classic rough set can only deal with discrete data; therefore, we must discrete the continuous attributes, which directly affects the subsequent attributes’ reduction and ensemble model construction.
As the data flow, we delete the old data synchronously and make sure that data appearing at the window are the need for processing, so as to realize the dynamic updating of the data set.
The data set of 12 attributes obtained by the reduction of rough set trains neural network classifiers, and integrating the results of each classifier to the ensemble.
On the last data chunk, concept drifting happened suddenly, and moreover, the accuracy after reduction is also higher than that before reduction.
Concluding Methods of traditional data mining face two major challenges of data stream: unlimited data and concept drifting.
Online since: February 2017
Authors: Lei Jiao Ge, Fu Jian Chi, Zhang You Xu, Yi Fang Su, Pan Zhang, Cong Liu, Shu Yan Yu, Rong Gang Ge
Fault Location Principle based on RS-IA Data Mining Model Overview.
It finds the laws from a large number of data by analyzing each data, and expresses them in the understandable form as far as possible [13, 15, 16].
Fault Location Based On RS-IA Data Mining Model Fault location based on the RS-IA data mining model mainly uses RS to extract domain knowledge, and obtains correlation rules of input and output vectors [14].
Fault Location of Distribution Network based on RS-IA Data Mining Model When there is a fault in some region of power distribution network, the fault location rules obtained by the RS-IA data mining model.
Sun, Data mining technology and its application on power system, Autom.
Online since: November 2010
Authors: Li Yang, Xin Zhang, Yan Zhang
The data driven refers to realizing the data predict that appraisal, dispatch, monitoring, diagnosis, decision-making and optimization of the system and so on each kind of expectation function by using the controlled system online and the off-line data [4].
Fig. 1 The multi-source information fusing system functional block diagram based on data driven Data Level Information Fusion.
Suppose there is n sensor to carry on measurement to some object X, each sensor measured data number is m, and then the sampled data matrix is:
Supposes the system input includes n sample, each sample Xi is composed of m data, produces the training sample data matrix (omitted).
References [1] Waltz E, Llinas J.Multisensor Data Fusion[M].
Online since: February 2012
Authors: De Yong Wang, Yan Liu, Mao Fa Jiang
The heat balance calculation of producing stainless steel crude melts by chromium ore smelting reduction in a 150 t converter is carried out by use of the empirical data and the calculation method of refining plain carbon steel in a converter, according to the blowing conditions of 185 t smelting reduction converter of No.4 steelmaking shop in Chiba Works of JFE Steel.
The settings of some smelting parameters refer to the empirical data of converter smelting.
Heat Balance Calculation The Required Raw Data of Calculation.
Basic raw data including the temperature of various charges and products, mean heat capacity of materials and chemical reaction heat effect at the smelting temperature are shown in Table 1-3.
out by use of the empirical data and calculation method of refining plain carbon steel in a converter, according to the blowing conditions of 185 t smelting reduction converter of No.4 steelmaking shop in Chiba Works of JFE Steel.
Online since: September 2014
Authors: Gui Fen Chen, Li Ma
Rough set theory has some advantages in the processing of data and eliminating redundant information and dealing with uncertain information,so widely used in data preprocessing, attribute reduction and so on.
Applications Based on Rough Sets and decision tree Experimental Data Acquisition The experimental data is from the Nong'an survey data, provided by Agricultural Technology Promotion Center.
Data Mining Based on Rough Set and Decision Tree combination.
Rough set attribute reduction algorithm requires that data is discrete data, according to the soil data characteristics, the Entropy / mdl discrete algorithm is carried to do the data processing.
Data Mining Results.In 1400 Records, correctly classified data is 1298, the others are not.
Online since: May 2014
Authors: Qiao Yan Li, Yan Yan Chen, Shao Yang Li
Introduction In our modern society, the volume and complexity of the collected data is growing rapidly.
In order to extract and describe the useful information in those data, the data mining emerges as a research area.
Data mining is a useful tool and method to solve the volume and complexity of the collected data.
Pawlak [1] in 1982, which is one of a new tools to dealing with vagueness and granularity in information systems, and it is able to find the useful information from volume and complexity of the collected data, moreover, it can guide people's decision by analyzed those useful information.
So is the N-reduction of fuzzy covering C.
Online since: September 2007
Authors: N. Ivanović, Milesa Srećković, N. Popović, V. Kusigerski, Lj. Vulićević, S. Vardić, Ž. Tomić
Reduction of Nanometric Magnetite Powder Lj.
The experimental data suggest that the reoxidation process proceeds in two stages, at various activation energies.
The mass loss (∆mT∼31%) measured after the reduction process at T=833 K is somewhat larger than the theoretically predicted one (∆mT=27,64%) for the reduction of magnetite to iron.
Assuming that the reaction is diffusion controlled, Jander's model was applied, and the experimental data were correlated with the equation (full lines in Fig. 4) [8]: [1-(1-α) 1/3 ] 2 = kτ
According to the experimental and calculated data, (see Fig.5), it seems that the (re)oxidation process has two distinct stages, at activation energies of 0.17 and 0.59 eV.
Online since: January 2014
Authors: Rui Xian Li
Traffic flow sampled data is noisy and chaotic time series.
Traditional ESN is only suitable for no noise or low-noise processing data, while the data with high noise prediction result is unsatisfactory[2].In interference of complex high-intensity noise component, simply optimization for ESN model does not completely solve the problem of prediction accuracy , while the use of data smoothing processing methods is often more direct and effective than optimal ESN model itself .The use of signal processing method filters out the data noise component, makes the data which is reduced noise have a higher signal noise ratio(SNA) , then provides the necessary conditions for high-precision prediction[3,4] .
The definition of goodness of fit, according to the References [6]: (2) Where is the original traffic flow data, is traffic flow data after noise reduction, is the sampling point .The Smaller is, the higher similarity to traffic flow before and after noise reduction is.
Multi-State Threshold Method Multi-state Threshold Method Which Is Based On Wavelet Transform In the Wavelet-based noise reduction algorithm, without any kind of soft-threshold method can be fit for all types of data.
Applicable threshold selection rules need us to fully consider characteristics of the denoising data.
Online since: December 2014
Authors: Xiao Xue Xing, Wei Wei Shang, Li Min Du
With the increasing of data in database systems, attribute reduction becomes more effective relative to the value reduction.
But the discernibility function simplification is a NP problem [5], so this algorithm is only suitable for with very small data set.
The dependence of attributes in rough set theory is the influence on decision rules in the current data, but it can’t reflect the prior knowledge of decision maker.
The experimental system chooses Visual C++6.0 as the development language, using Windows 2000 as the development environment, using six data sets from UCI standard data.
The data is processed after the missing information processing, and attributes are reduced using the above two algorithm respectively.
Online since: January 2011
Authors: Shi An, Xian Ye Ben, Jian Wang, Hai Yang Liu
We propose a novel method for data reduction in gait recognition, called Subblock Complete Two Dimensional Principal Component Analysis (SbC2DPCA).
This translates to data reduction with very minimal loss of information, as demonstrated by the remarkable recognition accuracy when subjects change clothing or have a backpack.
This is the second stage of data reduction in the process.
This balances recognition accuracy and data dimensionality.
C2DPCA is combined with blocking, to achieve further dimension reduction because of a large amount of data.
Showing 61 to 70 of 40196 items