We consider the task of finding out-of-class samples in tabular data, where little can be assumed on the structure of the data. In order to capture the structure of the samples of the single training class, we learn mappings that maximize the mutual information between each sample and the part that is masked out. The mappings are learned by employing a contrastive loss, which considers only one sample at a time. Once learned, we can score a test sample by measuring whether the learned mappings lead to a small contrastive loss using the masked parts of this sample. Our experiments show that our method leads by a sizable accuracy gap in comparison to the literature and that the same default rule of hyperparameters selection provides state-of-the-art results across benchmarks.
|State||Published - 2022|
|Event||10th International Conference on Learning Representations, ICLR 2022 - Virtual, Online|
Duration: 25 Apr 2022 → 29 Apr 2022
|Conference||10th International Conference on Learning Representations, ICLR 2022|
|Period||25/04/22 → 29/04/22|