Classification with Costly Features using Deep Reinforcement Learning
About
We study a classification problem where each feature can be acquired for a cost and the goal is to optimize a trade-off between the expected classification error and the feature cost. We revisit a former approach that has framed the problem as a sequential decision-making problem and solved it by Q-learning with a linear approximation, where individual actions are either requests for feature values or terminate the episode by providing a classification decision. On a set of eight problems, we demonstrate that by replacing the linear approximation with neural networks the approach becomes comparable to the state-of-the-art algorithms developed specifically for this problem. The approach is flexible, as it can be improved with any new reinforcement learning enhancement, it allows inclusion of pre-trained high-performance classifier, and unlike prior art, its performance is robust across all evaluated datasets.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Classification | WINE (test) | Accuracy83.33 | 29 | |
| Tabular Classification | Cirrhosis (test) | Accuracy74.6 | 14 | |
| Tabular Classification with Feature Selection | Heart | Accuracy70.66 | 14 | |
| Tabular Classification | Yeast (test) | Accuracy51.37 | 14 | |
| Tabular Classification with Feature Selection | Wine | Accuracy82.5 | 14 | |
| Tabular Classification with Feature Selection | Yeast | Accuracy0.3252 | 14 | |
| Tabular Classification | Diabetes (test) | Accuracy53.66 | 14 | |
| Tabular Classification | Heart (test) | Accuracy76.14 | 14 | |
| Tabular Classification with Feature Selection | Diabetes | Accuracy55.56 | 14 | |
| Tabular Classification with Feature Selection | Cirrhosis | Accuracy49.64 | 14 |