What is it about?
A common problem in meta-learning is establishing a (good) collection of meta-features that best represent the dataset properties[5]. In other words, the meta-learning recommendation's quality depends on the meta-features decision quality, and their ability to reflect the actual data challenges for the given dataset. Hence, the research question of this study is to what extent meta-features can describe the actual data difficulty without being affected by complex data challenges and thereby produced biased recommendation? According to literature, this question has not been given much attention but instead most of the works in this context focus on validating the meta-learning recommendation by evaluating the learning algorithms prediction performance (i.e., identifying correlations between meta-learning outputs and learning algorithm performance).
Featured Image
Photo by NASA on Unsplash
Why is it important?
To ensure the success of the meta-learning model since the learning algorithm selection decision is based on meta-feature.However, little attention has been paid to validating the meta-feature decisions in reflecting the actual data properties. In particular, if the meta-feature analysis is negatively affected by complex data characteristics, such as class overlap due to the distortion imposed by the noisy features at the decision boundary of the classes and thereby produces biased meta-learning recommendations that do not match the actual data characteristics (either by overestimating or underestimating the complexity).
Perspectives
Read the Original
This page is a summary of: Investigating the Performance of Data Complexity & Instance Hardness Measures as A Meta-Feature in Overlapping Classes Problem, August 2023, ACM (Association for Computing Machinery),
DOI: 10.1145/3616131.3616132.
You can read the full text:
Contributors
The following have contributed to this page