Wrapper Methods in Feature Selection: A Greedy Approach?
B. False
Wrapper methods in feature selection do not evaluate all possible combinations of features, even though they follow a greedy search approach.
Here's why:
- Greedy Search: Wrapper methods use a greedy search strategy, making locally optimal choices at each step with the hope of finding a globally optimal solution. * Computational Expense: Evaluating all possible feature combinations is computationally expensive, especially with a large number of features. * Subset Evaluation: Instead of exhaustive search, wrapper methods evaluate a subset of feature combinations, guided by the performance of a chosen machine learning algorithm. They start with an empty or full set and iteratively add or remove features based on the algorithm's performance on a hold-out validation set.
Examples of wrapper methods:
- Forward Selection: Starts with an empty set and adds one feature at a time.* Backward Elimination: Starts with all features and eliminates one at a time.* Recursive Feature Elimination: Recursively considers smaller and smaller sets of features.
While wrapper methods offer a more computationally efficient approach compared to exhaustive search, they don't guarantee finding the absolute best feature subset.
原文地址: https://www.cveoy.top/t/topic/R6p 著作权归作者所有。请勿转载和采集!