One-Shot Generalization and Open-Set Classification

Written by fewshot | Published 2024/08/27
Tech Story Tags: neural-networks | machine-learning-benchmarks | continual-learning-for-llm | class-agnostic-network | disentangled-learning | open-set-classification | class-incremental-learning | forward-and-backward-transfer

TLDRIn one-shot learning experiments, our model demonstrates strong generalization to unseen shapes, with high accuracy maintained even as new test classes are added. For open-set classification, our model detects novel shapes effectively by classifying inputs as unseen if they don't match existing exemplars significantly. This showcases the robustness and universal applicability of our equivariant learning framework.via the TL;DR App

Authors:

(1) Sebastian Dziadzio, University of Tübingen ([email protected]);

(2) Çagatay Yıldız, University of Tübingen;

(3) Gido M. van de Ven, KU Leuven;

(4) Tomasz Trzcinski, IDEAS NCBR, Warsaw University of Technology, Tooploox;

(5) Tinne Tuytelaars, KU Leuven;

(6) Matthias Bethge, University of Tübingen.

Table of Links

Abstract and 1. Introduction

2. Two problems with the current approach to class-incremental continual learning

3. Methods and 3.1. Infinite dSprites

3.2. Disentangled learning

4. Related work

4.1. Continual learning and 4.2. Benchmarking continual learning

5. Experiments

5.1. Regularization methods and 5.2. Replay-based methods

5.3. Do we need equivariance?

5.4. One-shot generalization and 5.5. Open-set classification

5.6. Online vs. offline

Conclusion, Acknowledgments and References

Supplementary Material

5.4. One-shot generalization

To evaluate whether the learned regression network can generalize to unseen classes, we perform a one-shot learning experiment. Here, the model had to normalize and classify transformed versions of shapes it had not previously encountered.

Since the returned class label depends on the exemplars in the buffer, we consider two variants of the experiment, corresponding to generalized and standard one-shot learning. In the first one, we keep the training exemplars in the buffer and add new ones. In the second, the buffer is cleared before including novel exemplars. We also introduce different numbers of test classes. The classification accuracies are presented in Table 1. As expected, keeping the training exemplars in the buffer and adding more test classes makes the task harder. Nevertheless, the accuracy stays remarkably high, showing that the equivariant network has learned a correct and universal mechanism that works even for previously unseen shapes. This is the essence of our framework.

5.5. Open-set classification

Next, we investigate how well our proposed framework can detect novel shapes. This differs from the one-shot generalization task because we do not add the exemplars corresponding to the novel shapes to the buffer. Instead of modifying the learning setup, we use a simple heuristic based on an empirical observation that our model can almost perfectly normalize any input—we classify the input image as unseen if we can’t find an exemplar that matches the normalized input significantly better than others.

This paper is available on arxiv under CC 4.0 license.


Written by fewshot | Spearheading research, publications, and advancements in few-shot learning, and redefining artificial intelligence.
Published by HackerNoon on 2024/08/27