Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning Hyper Label Model for Programmatic Weak Supervision

About

To reduce the human annotation efforts, the programmatic weak supervision (PWS) paradigm abstracts weak supervision sources as labeling functions (LFs) and involves a label model to aggregate the output of multiple LFs to produce training labels. Most existing label models require a parameter learning step for each dataset. In this work, we present a hyper label model that (once learned) infers the ground-truth labels for each dataset in a single forward pass without dataset-specific parameter learning. The hyper label model approximates an optimal analytical (yet computationally intractable) solution of the ground-truth labels. We train the model on synthetic data generated in the way that ensures the model approximates the analytical optimal solution, and build the model upon Graph Neural Network (GNN) to ensure the model prediction being invariant (or equivariant) to the permutation of LFs (or data points). On 14 real-world datasets, our hyper label model outperforms the best existing methods in both accuracy (by 1.4 points on average) and efficiency (by six times on average). Our code is available at https://github.com/wurenzhi/hyper_label_model

Renzhi Wu, Shen-En Chen, Jieyu Zhang, Xu Chu• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet 1.0 (val)
Accuracy56.21
48
Unsupervised Ensemble LearningCSGO
Accuracy87.08
13
Unsupervised Ensemble LearningMicroAgg2
Accuracy62.93
13
Unsupervised Ensemble LearningPetFinder
Accuracy79.03
13
Unsupervised Ensemble LearningGesturePhsm
Accuracy66.25
13
Unsupervised Ensemble LearningTree3k
Accuracy94.65
13
Unsupervised Ensemble LearningMnistE
Accuracy85.21
13
Unsupervised Ensemble LearningEyeMovem
Accuracy71.09
13
Unsupervised Ensemble LearningArtiChars
Accuracy79.06
13
ClassificationAmpData Expert
Accuracy60.28
8
Showing 10 of 15 rows

Other info

Follow for update