"Krung Sinapiromsaran"
"1/24/2015"
Scientific method entails formulation of hypotheses from observed facts followed by deductions and verification repeated in a cyclical process.
Facts are observations which are taken to be true while Hypothesis is a tentative conjecture regarding the phenomenon under consideration.
Deductions are made out of the hypotheses through logical arguments which in turn are verified through objective methods.
However, many physical processes, biological phenomena are characterised by variation and uncertainty.
Experiments when repeated under similar conditions need not yield identical results, being subjected to fluctuations of random nature.
Observations on the complete set of individuals in the population are impossible only a sample is considered.
See a list of my current publications
Most machine learning algorithms optimize the overall classification accuracy
SMOTE: Synthetic Minority Over-sampling Technique (2002)
Borderline-SMOTE (2005)
Safe-Level-SMOTE (2009) Chumphol
SMOUTE (2010) Panote
Minority-based Decision tree (July 2011) Kesinee
MUTE (December 2011) Chumphol
DBSMOTE (2012) Chumphol
Safe Level Graph for Majority Under-sampling Techniques (2014)
Safe-Level-SMOTE: Safe-Level-Synthetic Minority Over-sampling Technique for handling the class imbalanced problem
The 13th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD 2009), 27-30 April 2009, Bangkok, Thailand
| Region | Definition |
|---|---|
| Noise | \( n \) = \( k \) |
| Borderline | \( \frac{k}{2} \leq n < k \) |
| Safe | \( 0 \leq n < \frac{k}{2} \) |
\( n \) is the number of negative instances among the \( k \) nearest neighbors
| Name | Instance | Attribute | Positive | Negative | %Minority |
|---|---|---|---|---|---|
| Satimage | 6,435 | 37 | 626 | 5,809 | 9.73 |
| Haberman | 306 | 4 | 81 | 225 | 26.47 |
Nice paper, although there are several issues that avoid a better rating:
The main complaint about the poor experimental results .Since, the paper does not provide any theoretical framework and the results are primarily based on empirical evidence, it is important that the authors demonstrate the superiority of the algorithm using atleast 5-6 standard datasets.
There are many datasets that contain the minority data less than 10%. I encourage that the authors carefully look into papers on SMOTE and others that they cite, to get the appropriate datasets. Hence, I recommend a reject for this paper.
Overall the paper is well written. Regarding the over sampling, i am wondering if you keep the original data to test the techniques the result is the same or not? When you use 10-fold cross validation on new dataset for evaluation, the oversampled positive instances are included in performance testing. Because of the increase of the positive instances, the performance will be changed definitely. Have you try to test the performance only using original data? What is result?
The English needs to be improved. There are many typos such as “positive stances” (it should be read “positive instances”). The Figs are a bit small for reading.
Fig 3-5 are not very convincing. There does not seem to be much improvement.
Often in practise I find a random forest is good at under-represented positives. Is it better to use SMOTE to pre-process or just rely on RF?
Would confidence levels around the performance wipe out any differences in performance?
It would be better in the figure legends if ORG did not show the line as there is no line - was confusing to start with.
Requires a good proof read to fix grammar.
The results are not very convincing.
For a particular dataset, how well does this new method perform comparing with existing classifiers?
For a particular classifier, how effective does this new method perform on various datasets?
For a future unknown dataset, would you recommend this methodology to be used and why?
A t-test can be used to determine if two sets of data are significantly different from each other.
It can be used to answer the question of “For a particular dataset, how well does this new method perform comparing with existing classifiers?”
By performing repeated experiments of sampling the same datasets many times, it is easy to compute (paired) t-statistics and p-value.
The Wilcoxon singed-rank test is a non-parametric statistical hypothesis test used when comparing two related samples, matched samples, or repeated measurments on a single sample to assess whether their population mean ranks differ. Wikipedia
It can be used to answer the question of “For a particular classifier, how effective does this new method perform on various datasets?”
By performing repeated experiments on various datasets, the normality assumption can not be used.
Varying datasets and classifiers, the hypothesis testing can not rely on normality assumption so the non-parametric must be used.
It can be used to answer “For a future unknown dataset, would you recommend this methodology to be used and why?”