Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Deep Direct Likelihood Knockoffs

About

Predictive modeling often uses black box machine learning methods, such as deep neural networks, to achieve state-of-the-art performance. In scientific domains, the scientist often wishes to discover which features are actually important for making the predictions. These discoveries may lead to costly follow-up experiments and as such it is important that the error rate on discoveries is not too high. Model-X knockoffs enable important features to be discovered with control of the FDR. However, knockoffs require rich generative models capable of accurately modeling the knockoff features while ensuring they obey the so-called "swap" property. We develop Deep Direct Likelihood Knockoffs (DDLK), which directly minimizes the KL divergence implied by the knockoff swap property. DDLK consists of two stages: it first maximizes the explicit likelihood of the features, then minimizes the KL divergence between the joint distribution of features and knockoffs and any swap between them. To ensure that the generated knockoffs are valid under any possible swap, DDLK uses the Gumbel-Softmax trick to optimize the knockoff generator under the worst-case swap. We find DDLK has higher power than baselines while controlling the false discovery rate on a variety of synthetic and real benchmarks including a task involving a large dataset from one of the epicenters of COVID-19.

Mukund Sudarshan, Wesley Tansey, Rajesh Ranganath• 2020

Related benchmarks

TaskDatasetResultRank
Metabolite IdentificationIBD dataset (Case study)
Referenced Metabolites Count17
5
Feature SelectionGaussian mixture p=100 (n=200)
FDR Mean0.772
5
Feature SelectionGaussian mixture n=2000 p=100
FDR Mean0.725
5
Showing 3 of 3 rows

Other info

Follow for update