News & Events


Mini calevent portlet

Upcoming events in NorwAI

Previous events

Banner Previous events

Webinar series

Webinar Series on AI Research and Innovation

Previous webinars in the NorwAI&NAIL Friday Webinar Series

Previous webinars in the NorwAI&NAIL Friday Webinar Series

Speaker Topic

Simen Eide,


Simen Eide is an industrial PhD candidate at the University of Oslo and focusing on recommender systems, decision making and model uncertainty. Prior to the PhD, he worked as a practitioner building recommender systems and other machine learning products for the marketplace. He has a masters in mathematics with focus in mathematical finance and also worked on building portfolio risk systems for the Oslo and Swiss exchange.



Recommender systems, bandits and bayesian neural networks


Internet platforms consist of millions or billions of different items that users can consume. To help users navigate in this landscape, recommender systems have become an important component in many platforms.


The aim of a recommender system is to suggest the most relevant content on the platform to the user based on previous interactions the user has done with the platform.


A model being used in recommender systems is faced with multiple sources of uncertainty: There are limited interactions per user, the signals a user makes may be noisy and not always reflecting her preferences, and new items may be introduced to the platform giving few signals on these items as well. This talk will focus on model uncertainty and decision making in the recommender systems domain, with focus on the Norwegian marketplace We will discuss various ways to quantify, reduce and exploit these uncertainties through the use of bayesian neural networks, hierarchical priors and different recommender strategies

Speaker Topic

Professor Mark Keane (University College of Dublin)



Augmenting the Weather: Using counterfactuals to deal with dataset drift caused by climate change


In recent years, counterfactuals has become very popular for explaining the predictions of black-box AI systems. For example, if you are refused a loan by an AI and ask “why”, a counterfactual explanation might tell you, “well, if you asked for a smaller loan, then you would have been granted the loan". These counterfactuals are generated by methods that perform perturbations of the feature values of the original query (e.g., we perturb the value of the loan), and are typically synthetic data-points that did not originally occur in the dataset. 

This aspect of counterfactual methods prompted us to consider whether they might also work for data augmentation; that is, the supplementation of a dataset with generated (rather than actual) data-points. Data augmentation is important to Deep Learning models (where there may be a scarcity of data) and to prediction problems where there is data-drift or concept-drift (because the data is changing over time). A classic case of the latter is climate change. As our climate changes, past data on the weather is drifting towards more (and often more extreme) climate-disrupted events. 

If we are to predict phenomena using climate data, we need to be able to track these changes. We report work we have done using counterfactual methods to augment data to improve prediction in the face of such drifting. We also show that this method seems to generalise to imbalanced datasets and does better than very popular data augmentation methods (such as SMOTE). (Joint work with Mohammed Temraz & Barry Smyth)

Speaker Topic
  • Anastasios Lekkas (Associate Professor)
  • Inga Strümke (Postdoctoral fellow)
  • Vilde Gjærum  (PhD student)
  • Sindre Remman (PhD student)

Explainable AI


The EXAIGON project (2020-2024) delivers research and competence building on Explainable AI, including algorithm design and human-machine co-behaviour, to meet the society’s and industry’s standards for deployment of trustworthy AI systems in social environments and business-critical applications.

Speaker Topic

Jurica Šprem, GE Healthcare


Jurica Šprem has a Bachelor's degree in Computing from the University of Zagreb and a Master’s degree in ICT with a focus on signal processing, also from University of Zagreb.

In his PhD, he worked with enhanced cardiovascular risk prediction by machine learning at the Image Science Institute, UMC Utrecht. After obtaining his PhD in 2019, Jurica joined GE Healthcare in Oslo as AI Tech Lead with a focus on combining AI with cardiac ultrasound. He is currently a Product Owner of AI within cardiac ultrasound where he continues to follow his interests in combining machine learning and AI together with medical imaging.

Making Healthcare More Human With AI


We have experienced significant changes in our everyday life in the past years. We see new technologies emerge almost on daily basis, surrounding us with different tools and solutions with the goal to improve our lifestyle. Artificial intelligence (AI) is one of such technologies that has been entangled in almost all aspects of our daily lives. 

But can AI help healthcare professionals do their jobs the way they always wanted to by providing them with time and tools to focus on what matters and build a more efficient and intelligent ecosystem for patient care? Here we discuss problems and solutions AI brings within healthcare, how the cardiac ultrasound division at GE has adopted AI, and how AI is making health care more human.

Speaker Topic

Armin Catovic, Schibsted


Armin Catovic graduated 2007 in Computer Science/Telecom Engineering double degree from Swinburne University, Melbourne, Australia. He worked for a number of startups from 2005 before joining Ericsson in 2008. He spent the next 13 years at Ericsson working in various roles including radio engineer, system tester, software engineer and a machine learning engineer, and in various countries - Australia, Indonesia, Bangladesh, Singapore, US and Sweden. He's been working as a data scientist at Schibsted since the beginning of 2021, focusing on natural language processing. He currently lives in Stockholm with his wife and two kids.

Machine Learning in Contextual Advertising


Contextual advertising is a form of targeted advertising where the content of an ad segment is directly correlated with the content of a news article or a web page. In this presentation we walk through our use of unsupervised topic models, in order to optimally map demand with our inventory. We discuss caveats and challenges when working with unsupervised models in production. We also look into future work combining machine learning and contextual advertising.

Speaker Topic

Sara Malacarne, Telenor & Massimiliano Ruocco, SINTEF & NTNU


Sara Malacarne is a Research Scientist in the Analytics and AI team at Telenor Research. Sara has PhD in pure math. Since joining Telenor Research in 2018, she has been developing an interests towards AI/DL methods for solving tasks in the time series domain for 4G/5G Telco data. She is a collaborator in the ML4ITS “Machine Learning for Irregular Time Series” project lead by Massimiliano Ruocco (PI).
Massimiliano Ruocco is a Senior Researcher at SINTEF Digital and Adj. Associate Professor at the Dept of Computer Science (IDI), at NTNU. 

Generative Adversarial Networks for Anomaly Detection on Telco multivariate time series


Anomaly detection is the process of identifying interesting events that deviate from the data’s “normal” behaviour and has many important applications to real case scenarios. In the telecommunications domain, efficient and accurate anomaly detection is vital to be able to continuously monitor the network infrastructure key performance indicators (KPIs) and alert for possible incidents in time.
Network KPIs are in the form of multivariate time series which, for costs reasons, are not labelled. The main challenges for performing anomaly detection on network data are the following: 1) it is an unsupervised learning problem, 2) temporal and feature-wise correlations have to be exploited in order to reduce false positives,  3) anomalies are not necessarily rare events in the data, 4) the data is high-dimensional.  
This work is a first attempt to simultaneously address the first three challenges listed above, with the use of a novel Generative Adversarial Network (GAN), called RegGAN. GANs present in the literature -- such as MAD-GAN, BeatGAN, TadGAN -- have serious drawbacks on highly contaminated data, that is, data with frequent abnormal events. Thus, RegGAN was specifically built to overcome this issue, and it has proven to be robust to contamination experiments performed on open benchmark datasets.

Speaker Topic

Ole Jacob Mengshoel , NTNU & Ritchie Lee, NASA Ames Research Center


Dr. Ole Jakob Mengshoel is a Professor in Artificial Intelligence at the Department of Computer Science (IDI) at the Norwegian University of Science and Technology (NTNU) in Trondheim, Norway. At NTNU, he is affiliated with the Norwegian Open AI Lab. Dr. Mengshoel has published over 100 articles and papers in peer-reviewed journals and conferences, and holds 4 U.S. patents. He holds a Ph.D. in Computer Science from the University of Illinois, Urbana-Champaign. His undergraduate degree is in Computer Science from the Norwegian Institute of Technology, Trondheim (now NTNU).

Ritchie Lee is a research scientist in the Robust Software Engineering (RSE) group at NASA Ames Research Center.
His research interests are in safety validation and testing, decision-making systems, machine learning, and controls.  Of particular interest is developing algorithmic tools for the design and analysis of safety-critical cyber-physical systems including aircraft collision avoidance systems, air traffic automation, autonomous ground vehicles, and unmanned aerial systems (UASs). Ritchie holds a Ph.D. degree in Electrical and Computer Engineering from Carnegie Mellon University, an M.S. degree in Aeronautics and Astronautics Engineering from Stanford University, and a B.S. degree in Electrical Engineering from University of Waterloo.

Developing Complex but Trustworthy Computing Systems with Artificial Intelligence


This talk centers on the development of complex and trustworthy computing systems with artificial intelligence (AI).  Clearly, one can focus on or mean different things when discussing "AI and trustworthiness."  A first possible meaning is the development of trustworthy complex computing systems, systems that use AI, likely along with other computational methods.  Here, the development method itself may or may not use AI. A second meaning is to use AI methods to develop trustworthy complex computing systems, systems that may or may not contain or use AI themselves. 

In this case, the use of AI during development of a complex engineered system is front and center. In this talk we discuss both types of approaches. A particular focus is on a method called adaptive stress testing (AST), which falls in the second category mentioned above. Using simulations, AST finds likely failure events of complex aerospace systems by means of reinforcement learning and other AI techniques.  The AST method has, for example, been used to validate next-generation aircraft anti-collision systems.  

The development and validation of complex engineered systems with AI and different degrees of autonomy, up to fully autonomus systems, are key in developing trustworthy and human-centric AI. As these complex engineered systems proliferate and interact with each other and humans, we consider trustworthiness to be an essential topic in both fundamental and applied AI.

Speaker Topic

Stein H. Danielsen


Stein H. Danielsen is Co-founder and Chief Solutions Officer at Cognite.

Towards smart and autonomous industry


The NorwAI Innovate conference, took place in Trondheim, October 20-21. The conference brought together AI enthusiasts to present great examples of innovation research and industrial excellence within AI. In addition to the conference, Cognite and NTNU hosted a hackathon for students, as a side program to the main event.

One week after the conference and hackathon days, we’re happy to invite you to an open webinar, Friday October 29th. In this webinar, you’ll get to learn about the hackathon challenge and the winner team will pitch their solution to it. After that, Stein H. Danielsen from Cognite will give a talk about the mission of Cognite to work towards smart and autonomous industry: 

When we envision technology in the future we often think about robots - and their ability to solve complex tasks that would normally only be possible by humans. For industrial companies, robots are an essential part of their digitalization efforts, as their core business revolves around physical assets. To avoid repetitive, boring, and dangerous tasks, it is necessary for computer systems to interact with the real world. 

Stein Danielsen has always been passionate about robots, and is now co-founder and CSO of Cognite. He will tell us what robots can already do today, and show how robots are being put to use in Cognite. Furthermore, Stein will discuss how Cognite develops human-like understanding of the industrial reality for robots, and how we can exploit robot’s super-human capabilities. Finally, Stein will tell us where he thinks we’re heading for the future.

Speaker Topic

Martin Tveten


Martin Tveten is a research scientist at the Norwegian Computing Center, and a former PhD student at the Department of Mathematics, University of Oslo, specialising in methods and algorithms for change and anomaly detection. Applied interests currently include real-time monitoring of IT and industrial system. 

Introduction to change detection


In this seminar, I will introduce some basic ideas underlying statistical change detection methods. Such methods are important for answering any questions of the form "has some statistical property of the data changed over time?" and "if so, when does the change(s) occur?". An important AI-related application of change detection methods is anomaly detection in streaming data, for example from sensor networks.
Both the offline and online version of the change detection problem will be considered. In the offline problem, the aim is to retrospectively estimate the points in time where some statistical properties of a time series change. In the online problem, streaming data is processed in real-time with the aim of detecting a change as quickly as possible. A few simple, real and simulated, data examples will guide the presentation throughout, as the focus is on giving an intuition for the general methodology. I will also briefly present my own research.

NorwAI Innovate Banner

The conference NorwAI Innovate was held for the very first time on 20-21 October 2021 in Trondheim.

Learn more on the conference website  

Trustworthy Complex and Intelligent Systems Webinar Series

This series is a collaboration between the European Safety, Reliability & Data Association (ESReDA), the ETH Zürich Chair of Intelligent Maintenance Systems, the ETH Risk Center, ETH Zürich-SUSTech Institute of Risk Analysis, Prediction and Management (Risks-X), the Norwegian Research Center for AI Innovation (NorwAI) and DNV.

Webinars explore the themes of trust, ethics and applications of AI and novel technology in complex and safety critical intelligent systems.

Previous webinars in the Trustworthy Complex and Intelligent Systems Webinar Series

Previous webinars in the Trustworthy Complex and Intelligent Systems Webinar Series


Speaker Topic

Maziar Raissi, University of Colorado Boulder


Maziar Raissi is currently an Assistant Professor of Applied Mathematics at the University of Colorado Boulder. Dr. Raissi received a Ph.D. in Applied Mathematics & Statistics, and Scientific Computations from University of Maryland College Park. He moved to Brown University to carry out postdoctoral research in the Division of Applied Mathematics. Dr. Raissi worked at NVIDIA in Silicon Valley for a little more than one year as a Senior Software Engineer before moving to Boulder. His expertise lies at the intersection of Probabilistic Machine Learning, Deep Learning, and Data Driven Scientific Computing. In particular, he has been actively involved in the design of learning machines that leverage the underlying physical laws and/or governing equations to extract patterns from high-dimensional data generated from experiments.

Data-Efficient Deep Learning using Physics-Informed Neural Networks 


A grand challenge with great opportunities is to develop a coherent framework that enables blending conservation laws, physical principles, and/or phenomenological behaviors expressed by differential equations with the vast data sets available in many fields of engineering, science, and technology. At the intersection of probabilistic machine learning, deep learning, and scientific computations, this work is pursuing the overall vision to establish promising new directions for harnessing the long-standing developments of classical methods in applied mathematics and mathematical physics to design learning machines with the ability to operate in complex domains without requiring large quantities of data. To materialize this vision, this work is exploring two complementary directions: 

  1. designing data-efficient learning machines capable of leveraging the underlying laws of physics, expressed by time dependent and non-linear differential equations, to extract patterns from high-dimensional data generated from experiments, and
  2. designing novel numerical algorithms that can seamlessly blend equations and noisy multi-fidelity data, infer latent quantities of interest (e.g., the solution to a differential equation), and naturally quantify uncertainty in computations.
Speaker Topic

Enrico Zio

CRC MINES ParisTech, France

Politecnico di Milano, Italy


Enrico Zio is full professor at the Centre for research on Risk and Crises (CRC) of Ecole de Mines, ParisTech, PSL University, France, full professor and President of the Alumni Association at Politecnico di Milano, Italy. 


His research focuses on the modeling of the failure-repair-maintenance behavior of components and complex systems, for the analysis of their reliability, maintainability, prognostics, safety, vulnerability, resilience and security characteristics, and on the development and use of Monte Carlo simulation methods, artificial intelligence techniques and optimization heuristics. 


In 2020, he has been awarded the prestigious Humboldt Research Award from the Alexander von Humboldt Foundation in Germany

Prognostics and Health Management for Condition-based and Predictive Maintenance: A Look In and a Look Out


A number of methods of Prognostics and Health Management (PHM) have been developed (and more are being developed) for use in diverse engineering applications. Yet, there are still a number of critical problems which impede full deployment of PHM and its benefits in practice. In this lecture, we look in some of these PHM challenges and look out to advancements for PHM deployment.


Speaker Topic

Øyvind Smogeli, CTO Zeabuz


Øyvind Smogeli is the CTO and co-founder of Zeabuz and an Adjunct Professor at NTNU. Øyvind received his PhD from NTNU in 2006, and has spent his career working on modeling, simulation, testing and verification, complex cyber-physical systems, and assurance of digital technologies. He has previously held positions as CTO, COO and CEO of Marine Cybernetics and as Research Program Director for Digital Assurance in DNV.

Zeabuz: Providing trust in a zero emission autonomous passenger ferry


Zeabuz is developing a new urban mobility system based on zero emission, autonomous passenger ferries. This endeavour comes with a huge trust challenge: How to prove the trustworthiness towards both passengers, authorities, municipalities, and mobility system operators? This trust challenge has many facets and many stakeholders. There is a need to balance safety and usefulness, balance technical safety and perceived safety, and balance the various stakeholder needs. To solve this, an assurance case is being established that can capture a wide range of claims and evidence in a structured way. This talk introduces the Zeabuz mobility concept, the autonomy architecture, then will focus on the many layers of trust and how to achieve this. The various components of the autonomy system and the simulation technology used to build trust in the autonomy are explained. An approach to build trust in the simulators through field experiments and regular operation will be presented. It will be shown how this all fits into the larger assurance case.

Speaker Topic

Martin Vechev, ETH Zürich


Martin Vechev is an Associate Professor at the Department of Computer Science, ETH Zurich. His work spans the intersection of machine learning and symbolic methods with applications to topics such as safety of artificial intelligence, quantum programming and security. He has co-founded 3 start-ups in the space of AI and security, the latest of which LatticeFlow aims to build and deploy trustworthy AI models.

Certified Deep Learning


In this talk I will discuss some of the latest progress we have made in the space of certifying AI systems, ranging from certification of deep neural networks to entire deep learning pipelines. In the process I will also discuss new neural architectures that are more amenable to certification as well as mathematical impossibility and complexity results that help guide new kinds of certified training methods.


Speaker Topic

Asun Lera St.Clair, DNV & André Ødegårdstuen, DNV


Dr. Asun Lera St.Clair, philosopher and sociologist, is Director of the Digital Assurance Program in DNV Group Research and Development and Senior Advisor for the Earth Sciences unit of the Barcelona Supercomputing Center (BSC). She has over 30 years of experience with designing and directing interdisciplinary user-driven and solutions-oriented research for global challenges in the interface of sustainable development and climate change, and more recently on the provision of trust on digital technologies and leveraging these for sustainable development.

André Ødegårdstuen works as a Senior Researcher at DNV where he focuses on the assurance of machine learning. André is active in the area of computer vision for drone surveys of industrial assets and monitoring of animal welfare. He has a background in physics and experience from the Point-of-Care diagnostic industry.

Trustworthy Industrial AI Systems


Trust in AI is a major concern of many societal stakeholders. These concerns relate to the delegation of decisions to technologies we do not fully understand, to the misuse of those technologies for illegal, unethical or rights violation purposes, or to the actual technical limitations of these cognitive technologies while we rush to deploy them into society. There is a fast-emerging debate around these questions, often named as responsible AI, AI ethics, or explainable AI. However, there is less discussion as to what should be considered a trustworthy AI system in industrial contexts. AI introduces complexity and creates digital risks.

While complexity in traditional mechanical systems is naturally limited by physical constraints and the laws of nature, complexity in integrated, software-driven systems – which do not necessarily follow well-established engineering principles – seems to easily exceed human comprehension.

In this presentation we will unpack the idea that the trustworthiness of an AI system is not very different from that of a leader or an expert to whom, or an organization to which, we delegate our authority to make decisions or provide recommendations to reach a particular goal. Similarly, we argue that AI systems should be subjected to the same quality assurance methods and principles we use for any other technology.


Speaker Topic

Peter Battaglia


Peter Battaglia is a research scientist at DeepMind working on approaches for reasoning about and interacting with complex systems.

Structured models of physics, objects, and scenes


This talk will describe various ways of using structured machine learning models for predicting complex physical dynamics, generating realistic objects, and constructing physical scenes. The key insight is that many systems can be represented as graphs with nodes connected by edges, which can be processed by graph neural networks and transformer-based models. By considering the underlying structure of the problem, and imposing inductive biases within our models that reflect them, we can often achieve more accurate, efficient, and generalizable performance than if we avoid using principled assumptions.

22 January 2021

Speaker Topic

Joseph Sifakis


Hear 2007 Turing Award winner Joseph Sifakis explain the challenges raised by the vision for trustworthy autonomous systems for the autonomous vehicle case and outline his hybrid design approach combining model-based and data-based techniques and seeking trade offs between performance and trustworthiness.

Why is it so hard to make self-driving cars? (Trustworthy autonomous systems)


Why is the problem of self-driving autonomous control so hard? Despite the enthusiastic involvement of big technological companies and investment of  billions of dollars, optimistic predictions about the realization of autonomous vehicles have yet to materialize.

AI Workshop at womENcourage Pre-Event

AI Workshop at womENcourage Pre-Event

About: The workshop will bring together researchers interested in Artificial Intelligence. It is co-located with the womENcourage satellite pre-event organised by the Better Balance in Informatics (BBI) and IDUN projects promoting Women in Computer Science.

When: Tuesday, 21st of September 2021

Where: Radisson Blu Hotel, Tromsø

More information