Learning from limited training data


Deep learning methods have the strength of steadily improving performance with more training data. In the real world, the availability of suitable training data will often be limited, and annotation of complex image data requires domain experts and is both costly and time consuming. To succeed in our innovation areas there is an absolute need to research new methodology to learn from limited and complex training data.


For real-life applications with complex images, training data will often be limited in the sense that annotations (labels) will often be sparse, even if the amount of acquired data may be vast. Annotations may also be incomplete or inconsistent (noisy) and they are generally made for other purposes than training machine learning algorithms, and thus may be less suited for that purpose. Moreover, the characteristics of complex image data are often very different from the standard images, making the current transfer learning go-to solution, based on pre-trained ImageNet models, infeasible because the image data of interest is statistically out-of-distribution with respect to the base model.

Main objective

To develop new deep learning methods to solve complex problems from limited training data.

Highlighted publications

On the Effects of Self-supervision and Contrastive Alignment in Deep Multi-view Clustering
December 19, 2023
We propose DeepMVC – a unified framework which includes many recent methods as instances.
Merging clustering into deep supervised neural network
June 8, 2023
Introducing the SuperCM technique to significantly improve classification results across various types of image data.
Addressing Distribution Shifts in Federated Learning for Enhanced Generalization Performance
June 4, 2023
Training and test data from different clients pose a challenge.