-
Notifications
You must be signed in to change notification settings - Fork 13
Edited Nearest Neighbor #115
Copy link
Copy link
Open
Labels
New Stepa new recipe stepa new recipe stepfeaturea feature request or enhancementa feature request or enhancement
Description
Feature Request
Add support for Edited Nearest Neighbor (ENN) under-sampling.
Description
Edited Nearest Neighbor is an under-sampling technique that removes samples whose class label differs from the majority of their k-nearest neighbors. This effectively removes noisy and borderline samples.
Algorithm
- For each sample in the dataset:
- Find its k-nearest neighbors
- If the majority of neighbors belong to a different class, mark the sample for removal
- Remove all marked samples
This is a cleaning method that removes misclassified or ambiguous samples, resulting in smoother decision boundaries.
Key Properties
- Removes noisy samples that are likely misclassified
- Cleans decision boundaries
- Can be applied to majority class only, or to all classes
- Often used as a preprocessing step before other methods
Relationship to Other Methods
- Extended by: Repeated Edited Nearest Neighbor (Repeated Edited Nearest Neighbor #173)
- Extended by: AllKNN (AllKNN #174)
- Related: Neighborhood Cleaning Rule (Neighborhood Cleaning Rule #116)
References
- Wilson, D. L. (1972). Asymptotic properties of nearest neighbor rules using edited data. IEEE Transactions on Systems, Man, and Cybernetics, (3), 408-421.
- https://imbalanced-learn.org/stable/references/generated/imblearn.under_sampling.EditedNearestNeighbours.html
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
New Stepa new recipe stepa new recipe stepfeaturea feature request or enhancementa feature request or enhancement