@@ -319,11 +319,11 @@ iteratively decide if a sample should be removed or not
319
319
3. Train a 1-KNN on `C `.
320
320
4. Go through the samples in set :math: `S`, sample by sample, and classify each one
321
321
using a 1 nearest neighbor rule (trained in 3).
322
- 5. If the sample is misclassified, add it to :math: `C`, otherwise do nothing .
322
+ 5. If the sample is misclassified, add it to :math: `C`, and go to step 6 .
323
323
6. Repeat steps 3 to 5 until all observations in `S ` have been examined.
324
324
325
325
The final dataset is `S `, containing all observations from the minority class and
326
- those from the majority that were miss-classified by the 1-KNN algorithms.
326
+ those from the majority that were miss-classified by the successive 1-KNN algorithms.
327
327
328
328
The :class: `CondensedNearestNeighbour ` can be used in the following manner::
329
329
@@ -334,7 +334,7 @@ The :class:`CondensedNearestNeighbour` can be used in the following manner::
334
334
[(0, 64), (1, 24), (2, 115)]
335
335
336
336
However, as illustrated in the figure below, :class: `CondensedNearestNeighbour `
337
- is sensitive to noise and will add noisy samples.
337
+ is sensitive to noise and may select noisy samples.
338
338
339
339
In an attempt to remove noisy observations, :class: `OneSidedSelection `
340
340
will first find the observations that are hard to classify, and then will use
@@ -345,8 +345,8 @@ will first find the observations that are hard to classify, and then will use
345
345
2. Add a sample from the targeted class (class to be under-sampled) in
346
346
:math: `C` and all other samples of this class in a set :math: `S`.
347
347
3. Train a 1-KNN on `C `.
348
- 4. Using a 1 nearest neighbor rule trained in 3, classify all samples
349
- in set :math: `S`.
348
+ 4. Using a 1 nearest neighbor rule trained in 3, classify all samples in
349
+ set :math: `S`.
350
350
5. Add all misclassified samples to :math: `C`.
351
351
6. Remove Tomek Links from :math: `C`.
352
352
0 commit comments