Banner Banner

Prof. Dr. Wojciech Samek

Icon

Fraunhofer Heinrich Hertz Institute
Department of Artificial Intelligence

Einsteinufer 37, D-10587 Berlin

© private

Prof. Dr. Wojciech Samek

Fellow

Fellow | BIFOLD

Professor | Technische Universität Berlin

Head of AI Department | Fraunhofer HHI

Wojciech Samek is a professor in the department of Electrical Engineering and Computer Science at the Technical University of Berlin and is jointly heading the department of Artificial Intelligence and the Explainable AI Group at Fraunhofer Heinrich Hertz Institute (HHI), Berlin, Germany. He studied computer science at Humboldt University of Berlin from 2004 to 2010, was visiting researcher at NASA Ames Research Center, CA, USA, and received the Ph.D. degree in machine learning from the Technische Universität Berlin in 2014. He is associated faculty at the ELLIS Unit Berlin and the DFG Graduate School BIOQIC, and member of the scientific advisory board of IDEAS NCBR. Furthermore, he is a senior editor of IEEE TNNLS, an editorial board member of Pattern Recognition, and an elected member of the IEEE MLSP Technical Committee. He is recipient of multiple best paper awards, including the 2020 Pattern Recognition Best Paper Award, and part of the expert group developing the ISO/IEC MPEG-17 NNC standard. He is the leading editor of the Springer book “Explainable AI: Interpreting, Explaining and Visualizing Deep Learning” and organizer of various special sessions, workshops and tutorials on topics such as explainable AI, neural network compression, and federated learning. He has co-authored more than 150 peer-reviewed journal and conference papers; some of them listed by Thomson Reuters as “Highly Cited Papers” (i.e., top 1%) in the field of Engineering.

2021 2020 Pattern Recognition Best Paper Award
2019 Best Paper Award at ICML Workshop on On-Device Machine Learning & Compact Deep Neural Network Representations
2019 Honorable Mention Award at IEEE AIVR
2016 Best Paper Award at ICML Workshop on Visualization for Deep Learning
2010 Scholarship of the DFG Research Training Group GRK 1589/1
2006 Scholarship of the German National Merit Foundation (“Studienstiftung des deutschen Volkes”)

  • Deep Learning
  • Interpretable Machine Learning
  • Model Compression
  • Computer Vision
  • Distributed Learning

Sören Becker, Johanna Vielhaben, Marcel Ackermann, Klaus-Robert Müller, Sebastian Lapuschkin, Wojciech Samek

AudioMNIST: Exploring Explainable Artificial Intelligence for audio analysis on a simple benchmark

December 19 , 2023
https://doi.org/10.1016/j.jfranklin.2023.11.038

Luca Longo, Mario Brcic, Federico Cabitza, Jaesik Choi, Roberto Confalonieri, J. Ser, Riccardo Guidotti, Yoichi Hayashi, Francisco Herrera, Andreas Holzinger, Richard Jiang, Hassan Khosravi, Freddy Lecue, Gianclaudio Malgieri, Andres Paez, Wojciech Samek, Johannes Schneider, Timo Speith, Simone Stumpf

Explainable Artificial Intelligence (XAI) 2.0: A Manifesto of Open Challenges and Interdisciplinary Research Directions

October 30 , 2023
https://arxiv.org/pdf/2310.19775.pdf

Reduan Achtibat, Maximilian Dreyer, Ilona Eisenbraun, Sebastian Bosse, Thomas Wiegand, Wojciech Samek, Sebastian Lapuschkin

From attribution maps to human-understandable explanations through Concept Relevance Propagation

September 20 , 2023
https://doi.org/10.1038/s42256-023-00711-8

Armin W. Thomas, Ulman Lindenberger, Wojciech Samek, Klaus-Robert Müller

Evaluating deep transfer learning for whole-brain cognitive decoding.

July 13 , 2023
https://doi.org/10.1016/j.jfranklin.2023.07.015

News
Explainable AI| Oct 04, 2023

AI - finally explainable to humans

As of today it remains difficult to understand how an AI system reaches its decisions. Scientists at the Fraunhofer Heinrich-Hertz-Institut (HHI) and the Berlin Institute for the Foundations of Learning and Data (BIFOLD) at TU Berlin have collaborated for many years to make AI explainable. In their new paper the researchers present Concept Relevance Propagation (CRP), a new method for explainable AI that can explain individual AI decisions as concepts understandable to humans. 

News
Machine Learning| Apr 28, 2022

“I want to move beyond purely ‘Explaining’ AI”

BIFOLD researcher Dr. Wojciech Samek has been appointed Professor of Machine Learning and Communications at TU Berlin with effect from 1 May 2022. Professor Samek heads the Department of Artificial Intelligence at the Fraunhofer Heinrich-Hertz-Institute. His goal is to further develop three areas: explainability and trustworthiness of artificial intelligence, the compression of neural networks, and so-called federated leaning. He aims to focus on the practical, methodological, and theoretical aspects of machine learning at the interface to other areas of application.

News
Machine Learning| Dec 15, 2021

Benchmarking Neural Network Explanations

Neural networks have found their way into many every day applications. During the past years they reached excellent performances on various largescale prediction tasks, ranging from computer vision, language processing or medical diagnosis. Even if in recent years AI research developed various techniques that uncover the decision-making process and detect so called “Clever Hans” predictors – there exists no ground truth-based evaluation framework for such explanation methods. BIFOLD researcher Dr. Wojciech Samek and his colleagues now established an Open Source ground truth framework, that provides a selective, controlled and realistic testbed for the evaluation of neural network explanations. The work will be published in Information Fusion.

Machine Learning| Mar 24, 2021

New workshop series “Trustworthy AI”

The AI for Good global summit is an all year digital event, featuring a weekly program of keynotes, workshops, interviews or Q&As. BIFOLD Fellow Dr. Wojciech Samek, head of department of Artificial Intelligence at Fraunhofer Heinrich Hertz Institute (HHI), is implementing a new online workshop series “Trustworthy AI” for this platform.

News
Machine Learning| Mar 10, 2021

Making the use of AI systems safe

BIFOLD Fellow Dr. Wojciech Samek and Luis Oala (Fraunhofer Heinrich Hertz Institute) together with Jan Macdonald and Maximilian März (TU Berlin) were honored with the award for “best scientific contribution” at this year’s medical imaging conference BVM. Their paper “Interval Neural Networks as Instability Detectors for Image Reconstructions” demonstrates how uncertainty quantification can be used to detect errors in deep learning models.

News
Machine Learning| Dec 09, 2020

BIFOLD PI Dr. Samek talks about explainable AI at NeurIPS 2020 social event

BIFOLD Principal Investigator Dr. Wojciech Samek (Fraunhofer HHI) talked about explainable and trustworthy AI at the “Decemberfest on Trustworthy AI Research” as part of the annual Conference on Neural Information Processing Systems (NeurIPS 2020). NeurIPS is a leading international conference on neural information processing systems, Machine Learning (ML) and their biological, technological, mathematical, and theoretical aspects.