Image:
ICML

Four Visual Intelligence-authored papers accepted for prestigious machine learning conference

Visual Intelligence will be well presented at ICML 2025—one of the leading international academic conference in machine learning—with four newly accepted research papers.

Four Visual Intelligence-authored papers accepted for prestigious machine learning conference

Visual Intelligence will be well presented at ICML 2025—one of the leading international academic conference in machine learning—with four newly accepted research papers.

By Petter Bjørklund, Communications Advisor at SFI Visual Intelligence

The International Conference on Machine Learning (ICML) attracts 7000 researchers from around the globe to share high impact research in machine learning and artificial intelligence (AI). This year's conference has an acceptance rate of around 26.9 per cent.

Centre Director Robert Jenssen is thrilled by how well represented Visual Intelligence will be at this year's ICML, which will be organized from July 13th - 19th.

“It is very important for us as a research centre in deep learning to contribute to the scientific progress of the field, laying the foundation for innovation and real-world impact. I am very proud of our researchers’ cutting-edge work to better extract information in neural networks, to better compress information, to enable interpretability and to leverage multimodality,” Jenssen says.

Learning representations without labels using contrasting learning

Professor and Principal Investigator (PI) Adín Ramírez Rivera is one of six Visual Intelligence researchers who got their paper accepted for ICML 2025. His paper—titled "Self-Organizing Visual Prototypes for Non-Parametric Representation Learning"—is about learning representations without labels using contrasting learning.

Adín Ramírez Rivera. Photo: UiO.

Contrasting existing approaches, Ramírez Rivera and co-authors take advantage of the relationship between data that they have seen during training and improve the comparisons—in the contrastive setup—by using relative information of absolute one—just like how current methods do.

"Our results show that using this relative information helps to learn better representations as evidenced by several tasks that we performed on image data. We outperform several existing methods and show that this proposal not only scales but also outperforms current learning setups," Ramírez Rivera says.

Novel general layer-wise quantization framework

The second VI paper—titled "Layer-wise Quantization for Quantized Optimistic Dual Averaging"—is authored by Associate Professor and PI Ali Ramezani-Kebrya. He proposes a general layer-wise quantization framework that takes into account the statistical heterogeneity across layers and an efficient solver for distributed variational inequalities.

Ali Ramezani-Kebrya. Photo: Private

Ramezani-Kebrya and co-authors establish tight variance and code-length bounds for layer-wise quantization, which generalize the bounds for global quantization frameworks.

"We empirically achieve up to a 150% speed-up over the baselines in end-to-end training time for training Wasserstein GAN on 12+ GPUs", Ramezani-Kebrya explains.

Framework for visually self-explainable document question answering

The third paper, titled "DocVXQA: Context-Aware Visual Explanations for Document Question Answering", proposes DocVXQA: a framework for visually self-explainable document question answering that produces accurate answers while generating visual heatmaps for interpretability.

The paper is authored by Postdoctoral Researcher and PI Changkyu Choi and other collaborators from Spain, France and Norway.

Changkyu Choi. Photo: UiT

"By encoding explainability principles as learning criteria, DocVXQA balances performance and trust through context-aware explanations," Choi says.

Novel multimodal variational autoencoder

Rogelio Andrade Mancisidor, a former PhD Candidate at UiT Machine Learning Group, is the main author of the fourth accepted VI paper—titled "Aggregation of Dependent Expert Distributions in Multimodal Variational Autoencoders". He is now an Associate Professor at BI and a Visual Intelligence collaborator.

His paper introduces the Consensus of Dependent Experts (CoDE) method, which models the dependence between single-modality distributions through their error of estimation and generalizes Product of Experts (PoE). The paper is co-authored by associate professors Michael Kampffmeyer and Shujian Yu, as well as Centre Director Robert Jenssen.

Rogelio Andrade Mancisidor. Photo: BI.

Multimodal Variational Autoencoders (VAEs) use the Product of Experts (PoE) or Mixture of Experts (MoE) methods to estimate consensus distributions by aggregating single-modality distributions and assuming independence for simplicity, which—according to Mancisidor—is an overly optimistic assumption. The CoDE method was proposed as a way of overcoming this limitation.

"We use CoDE to develop CoDE-VAE—a novel multimodal VAE that learns the contribution of each consensus distribution to the optimization. We argue that consensus distributions conditioned on more modalities or with relatively more information should contribute extra to the optimization," Mancisidor explains.

He says CoDE-VAE shows better performance in terms of balancing the trade-off between generative coherence and generative quality—as well as generating more precise log-likelihood estimations.

"In addition, our experiments support the hypothesis that data modalities are correlated, as they are simply data modalities on the same underlying object," Mancisidor adds.

Latest news

Anders Waldeland receives the Digital Trailblazer Award 2025

December 4, 2025

Congratulations to Senior Research Scientist Anders Waldeland, who was awarded the Digital Trailblazer Award 2025 at the Dig X Subsurface conference in Oslo, Norway.

sciencenorway.no: AI can help detect heart diseases more quickly

December 3, 2025

Researchers have developed an artificial intelligence that can automatically measure the heart's structure – both quickly and accurately (Popular science article on sciencenorway.no)

State Secretary Marianne Wilhelmsen visits SFI Visual Intelligence and UiT

November 26, 2025

State Secretary Marianne Wilhelmsen visited UiT The Arctic University of Norway to learn more about SFI Visual Intelligence and UiT's AI initiatives in education and research.

TV2.no: Sier Elon Musk er smartere enn Leonardo da Vinci

November 25, 2025

KI-chatboten Grok har fortalt brukere at verdens rikeste mann er både smartere og sprekere enn noen andre i verden – inkludert basketballstjernen LeBron James og Leonardo da Vinci (Norwegian news article on tv2.no)

Successful science communication workshop at Skibotn

November 21, 2025

The Visual Intelligence Graduate School gathered our early career researchers for a 3-Day Science Communication workshop at Skibotn field station outside of Tromsø, Norway.

uit.no: UiT og Aker Nscale sammen om storsatsing på kunstig intelligens

November 19, 2025

Onsdag inngikk Aker Nscale og UiT Norges arktiske universitet en ti-årig samarbeidsavtale for å utvikle og styrke kompetansemiljøene for kunstig intelligens i Narvik og Nord-Norge. Aker Nscale garanterer for 100 millioner kroner i avtaleperioden (news story on uit.no)

Two fruitful days at The Alan Turing Institute's headquarters

November 17, 2025

Centre Director Robert Jenssen and PhD Candidate Lars Uebbing had two fruitful days together with researchers at The Alan Turing Institute's headquarters in London

Anders Waldeland nominated for the Digital Trailblazer 2025 Award

November 12, 2025

Senior Research Scientist Anders Waldeland is nominated for the Digital Trailblazer 2025 Award. The winner is announced at the Dig X Subsurface conference in Oslo, Norway in December.

AI can help detect heart diseases more quickly

November 7, 2025

Visual Intelligence researchers have developed an AI to automatically measure the heart's structure – both quickly and accurately. They believe it can help doctors detect and treat cardiovascular diseases faster.

How can PET and AI help detect prostate cancer earlier?

November 5, 2025

Samuel Kuttner and Elin Kile presented research on PET and artificial intelligence at evening seminar on early detection of prostate cancer organized by the Norwegian Prostate Cancer Assocation.

Visual Intelligence represented at Svarte Natta 2025

October 29, 2025

Centre Director Robert Jenssen represented Visual Intelligence at Svarte Natta 2025 – North Norway's journalist and media conference organized by the Norwegian Union of Journalists.

My Research Stay at Visual Intelligence: Aitor Sánchez

October 5, 2025

Aitor Sánchez is a PhD candidate at the Intelligent Systems Group of the University of the Basque Country in Spain. He visited Visual Intelligence in Tromsø from March to June 2025.