Limited training data

Visual Intelligence aims to develop new deep learning models that solve problems involving complex images from limited training data.

Motivation

The performance of deep learning methods steadily improves with more training data. However, the availability of suitable training data is often limited. Additionally, labelling complex image data requires domain experts and is both costly and time-consuming.

This research challenge is heavily stressed by a majority of our user partners as an immediate need. To succeed in our innovation areas, it is absolutely necessary to research new methodology which learn from limited and complex training data.

Solving research challenges through new deep learning methodology

Methods which exploit weak, noisy and incompletely labelled data, be it through semi-supervised or semi-supervised approaches, make up a significant portion of our portfolio. Examples include the following:

• A self-supervised approach for content-based image retrieval of CT liver images.

• Explainable marine image analysis methods validated on multiple marine datasets, such as multi-frequency echosounder data and aerial imagery of sea mammals captured by drones.

• A self-supervised method for automatically detecting and classifying microfossils.

• Methods for automatic building change detection in aerial images based on self-supervised learning.

These methods represent time-effective and cost-effective approaches which make deep learning models less reliant on large data samples and labeled data. These improve the models’ efficiency and ability to generalize, making them more applicable in real-world settings.

Highlighted publications

Modular Superpixel Tokenization in Vision Transformers

August 28, 2024
By
Marius Aasan, Odd Kolbjørnsen, Anne Schistad Solberg, Adín Ramirez Rivera

Reinventing Self-Supervised Learning: The Magic of Memory in AI Training

July 29, 2024
By
Thalles Silva, Helio Pedrini, Adı́n Ramı́rez Rivera

Other publications

SPoT: Subpixel Placement of Tokens in Vision Transformers

By authors:

Martine Hjelkrem-Tan, Marius Aasan, Gabriel Y. Arteaga, and Adín Ramírez Rivera

Published in:

Workshop on Efficient Computing under Limited Resources: Visual Computing (ICCV 2025), Oct 19 – 23th, 2025, Honolulu, Hawai'i

on

October 19, 2025

Low-Rank Adaptations for increased Generalization in Foundation Model features

By authors:

Vilde Schulerud Bøe, Andreas Kleppe, Sebastian Foersch, Daniel-Christoph Wagner, Lill-Tove Rasmussen Busund, Adín Ramírez Rivera

Published in:

MICCAI Workshop on Computational Pathology with Multimodal Data (COMPAYL), DAEJEON, South Korea, 2025

on

September 27, 2025

Aggregation of Dependent Expert Distributions in Multimodal Variational Autoencoders

By authors:

Rogelio A Mancisidor, Robert Jenssen, Shujian Yu, Michael Kampffmeyer

Published in:

International Conference on Machine Learning (ICLM) 2025

on

August 13, 2025

Self-Organizing Visual Prototypes for Non-Parametric Representation Learning

By authors:

Thalles Silva, Helio Pedrini and Adín Ramírez Rivera

Published in:

Forty-Second International Conference on Machine Learning (ICML), Vancouver, Canada 13-19 July, 2025

on

July 13, 2025

Pixel-Level Predictions with Embedded Lookup Tables

By authors:

Marius Aasan, Adín Ramírez Rivera

Published in:

Proceedings of the Symposium of the Norwegian AI Society 2025, CEUR Workshop Proceedings ( ISSN 1613-0073)

on

June 17, 2025