Summary and Contributions: The paper introduces a technique for out-of-distribution (OOD) detection based on recent advances in contrastive self-supervised learning. The idea is to contrast an image against domain-shifted augmentations (mainly, rotations) of the same image. The paper leverages this simple technique in learning a scoring function for detecting OOD samples and shows very good results on standard OOD benchmarks.
Strengths: * The approach is a sound and simple adaptation of the well understood contrastive learning method. * The paper provides extensive empirical results on several OOD benchmarks, with very good numbers.
Weaknesses: * While experimental results seem strong, the current format raises some concerns regarding the fairness of comparison to prior work. Empirical evidence would be stronger if results from  are compared to the results of this paper. In addition, results were not directly compared to GOAD , and it is not clear the comparison is fair since results in  are better than the reported ones. Finally, it's always great to provide standard deviation from multiple runs, which is missing from all tables. * I find the choice of rotation as a distribution shifting augmentation a little unconvincing. While it might work for well curated datasets such CIFAR10 and ImageNet, this approach will not suffice for real datasets. * Several key choices are not well motivated, such as using the norm in the scoring function. References:  Hendrycks, Dan, et al. "Using self-supervised learning can improve model robustness and uncertainty." Advances in Neural Information Processing Systems. 2019.  Bergman, Liron, and Yedid Hoshen. "Classification-based anomaly detection for general data." arXiv preprint arXiv:2005.02359 (2020).
Correctness: More effort needs to be done in ensuring fairness of comparison to prior methods.
Clarity: The paper is easy to follow. I would prefer including discussion of some key modeling choices in the main text instead of moving it to the appendix.
Relation to Prior Work: Relationship to related work, especially  in references above, needs to be more emphasized.
Additional Feedback: Post rebuttal ------------------- Thanks for addressing my concerns about experimental results. I would like to apologize for misreading your results and I agree that your comparison with references  and  are solid. As a result, I have raised my score.
Summary and Contributions: The paper presents a simple yet effective method to detect OOD samples: some data transformations, seen as harmful to learn feature representations with contrastive learning, are useful to "augment" the set of OOD samples and can be used as negatives in a contrastive loss. The paper presents a new formulation of the contrastive loss, useful for detecting OOD samples, as shown in the robust experimental section. --- After rebuttal --- I have read the author's response and I maintain my score.
Strengths: The paper is well motivated and the contributions are very clear from the beginning. All choices in the methodology are properly motivated. Moreover, the experimental section presents strong results on different setups, datasets, and the ablation study reveals many insights. The main contribution in this paper can provide key insights for future work along these lines.
Weaknesses: I do no think the paper presents any important weakness, as the scope of the work is clear, methodology seems sound and experiments are robust.
Correctness: The empirical methodology seems correct.
Clarity: The paper is well written and clear.
Relation to Prior Work: The paper does a good job on stating the differences between the presented method and existing techniques. Moreover, it correctly cites the work on top of which authors have build upon and how their work differs on it.
Additional Feedback: In l.108-118 it seems as if the only transformation used for positive samples in the contrastive loss is the identity function. Later in l.173-175 it specifies which were the used transformations that are used to generate the positive samples. This was a bit confusing, perhaps l.108-118 should clarify that there is other transformations that, applied to an image, can be considered as positive samples? The statement in l116-119, could use a reference to the specific experiments in which they are based on. In l.207, the authors mention "notable performance in the cases when prior methods often fail" regarding Table 2. What does it mean to fail here? I agree the proposed method is better than the other, but at which threshold are you determining that the other methods fail? Maybe this sentence could be rephrased. In l.241-246, Table 6 is referenced, but it seems that the comments are about Table 5.
Summary and Contributions: The paper proposes a contrastive learning-based out-of-distribution detection method by using distributionally-shifted augmentation to construct training example pairs. The main contributions of this paper include a new contrastive learning loss and a novel score function to evaluate the outlierness of OOD samples.
Strengths: 1. The authors found that some augmentations on images can be useful for OOD detection and they proposed the so-called con-SI loss to learn the discriminative representation of in-distribution examples, which is quite novel. 2. Although there is no theoretical guarantee that the proposed score function, it does make sense to distinguish OOD examples from inliers. 3. The experiments are comprehensive, the authors tested the proposed method in various settings and the results are convincible.
Weaknesses: 1. Despite the strengths as mentioned before, the overall novelty is not enough for NeurIPS. The proposed method is highly related to contrastive learning and simCLR, so it can be seen as an incremental research. 2. The proposed method is limited to image data, as the transformation augmentation cannot applied to other types of data, such as temporal data. So I suggest the author add some keywords such as visual or image to the paper title. 3. The paper does not clearly explain why these augmentations are helpful for OOD detection. 4. It is better to report the performance variance in each setting.
Correctness: This paper is technical sound, the formulation and the notations are clear to me and the method is easy to follow.
Clarity: This paper is well-organized and well-written.
Relation to Prior Work: Yes, the authors discussed the differences between their work and major previous works in introduction.
Additional Feedback: Two additional questions to the authors: 1. Is the resnet-18 used in this paper pretrained or trained for the task from scratch? 2. As the discrimination scores for OOD examples are various among different classes, how do you set the threshold in practical scenarios.
Summary and Contributions: This paper proposes "Contrasting Shifted Instances (CSI)" as an approach of novelty detection. The proposed approach is a simple modification of existing contrastive learning approaches. The main idea of the paper is that in addition to contrasting the samples with other samples, we can contrast samples with augmentations of themselves. This is unlike existing approaches which force augmentations of the same sample to be close to each other. The authors also propose a score which characterizes the novelty of a test sample. The authors conduct extensive experiments to validate the proposed approach.
Strengths: Apart from a few minor short-comings, the paper is very-well written. The authors have clearly and concisely explained their approach. There is extensive experimental validation of the proposed approach. A major strength of the paper is the simplicity of the proposed approach.
Weaknesses: The major weakness of the paper is that the authors do not relate their work with recent work on the same topic. I think a section relating the proposed work to prior work will make the paper much better. In addition, the authors should answer the following questions in the rebuttal/re-submission: 1. The authors have mentioned that con-SI does not improve the representation for standard classification. However, it is not clear whether the proposed approach harms standard classification. The authors should clarify whether the performance for standard classification changes a lot with con-SI. If the performance degrades, by how much? The authors should also discuss if there are ways to avoid such degradations in the performance. 2. Did the authors employ a principled way of selecting the set of transformations S? Currently, it's not clear how did the authors select the number of transformations to use and also which transformations to include in S. They have conducted ablation studies to understand the effectiveness of each transformation separately. However, can the authors discuss more principled ways of searching for elements of S? I think there might be some adversarial methods for selecting the best transformations.
Correctness: Yes. I have verified some of the mathematical formulation in the paper and did not find any major issues.
Clarity: Yes. The authors have clearly explained the approach and experiments.
Relation to Prior Work: Not in depth. This is a major weakness of the paper.
Additional Feedback: No changes after the rebuttal. I think this is a good paper.