Since an increasing number of scientists at the University of Freiburg engage in artificial intelligence research, we have decided to establish a network, particularly for Ph.D. students and PostDocs, to create a collaborative environment beyond faculty borders and foster exchange concerning various topics in AI research.
The University of Freiburg identifies data analysis and artificial intelligence as one of the university-wide profile fields. The profile field with its speakers, Prof. Dr. Harald Binder, Prof. Dr. Thomas Brox and Prof. Dr. Frank Hutter has set itself the goal of deepening interdisciplinary cooperation in the field of artificial intelligence at our university.
To ensure this cooperation also at the level of young scientists, we plan to bring together Ph.D. students and PostDocs from all faculties and institutions of the University of Freiburg to establish a highly interdisciplinary and diverse network of researchers to benefit from each other mutually. We are convinced it will be very valuable to know the approaches, issues, and techniques of people in other AI fields, invite exciting keynote speakers, and organize workshops based on common interests.
June 6th and June 7th 2023
10:00am – 5:00pm
HSII, Albertstraße 23b
Prof. Dr. Ribana Roscher (University of Bonn)
Title: Generating the Unseen and Explaining the Seen
Abstract: Deep generative models and explainable machine learning are two emerging areas of data science that we can use to address current challenges in agricultural and environmental sciences. Deep generative models are neural networks that are capable of learning complex data distributions. In general, they can be used for a variety of applications, such as anomaly detection or prediction. Explainable machine learning, which analyzes the decision-making process of machine learning methods in more detail, is used whenever an explanation for the result is required in addition to the result. This can be done for various reasons, e.g., to increase confidence in the outcome or to derive new scientific knowledge that can be inferred from patterns in the decision process of the machine learning model. This talk addresses methods from both areas and provides examples of their application in agricultural and environmental sciences.
Leon Sixt (Free University of Berlin)
Title: Explainable AI: History, Progress, and Future Directions
Abstarct: Explainable Artificial Intelligence (XAI) has become an essential aspect of AI research as the need for understanding and interpreting complex models grows in parallel with their capabilities. In this talk, I will explore the history since the 70s and developments in XAI, discuss the limitations of current methods, and present my own work in evaluating explainability methods. I will begin by tracing the origins of explanations in AI and highlight the key advancements in the last years. I will then delve into the current state of the field, including established methods like Integrated Gradients, LIME, and SHAP, and their limitations.
My research focuses on the evaluation of explainability methods, covering topics such as the examination of saliency maps and why some methods fail the sanity check, the evaluation of Deep Taylor Decomposition, and a user study on counterfactual explanations. I will also introduce the DNNR model, a simple and interpretable extension of KNN-regression. Lastly, I will discuss potential future research directions, including the evaluation of explanations and the development of explanations for large language models. This talk aims to provide a comprehensive overview of the challenges and progress in Explainable AI.
Dr. Stephan Alaniz (University of Tübingen)
Title: Explanations through Abstraction and Natural Language
Abstarct:
Many real-world application now and in the future will require us to better understand the decision process of deep neural networks. In our research, we explore a diverse set of means of explainability. This talk deals with two such approaches: abstractions and natural language. In one part of the talk, I will discuss how we can learn abstractions of our data that can help both us, the humans, to understand the data semantically as well as the model in downstream tasks. In the other part, I will present how we can explain the internal decision making process of neural networks using natural language either in the form of semantic attributes or with the help of a language model to obtain full descriptions.
Time | Plan |
---|---|
10:00-10:05 | Opening |
10:05-11:00 | Talk I: Dr. Stephan Alaniz |
11:00-12:00 | Talk II: Leon Sixt |
12:00-13:30 | Lunch break |
13:00-14:30 | Speed networking |
14:30-15:30 | Short Talks |
15:30-16:00 | Break |
16:00-17:00 | Talk III: Prof. Dr. Ribana Roscher |
17:00-Inf | Eating/Drinking together |
Machine learning (ML) is essential to data-driven research and can be used to make predictions for new data, to crunch data or to perform data analysis. However, applying ML can be quite complex as it comes with many design decisions, each impacting the final performance. For example, to build a predictive pipeline, one must construct appropriate features, design workflows, pick the right algorithms, and tune their hyperparameters. Automated Machine Learning (AutoML) aims to automate this process and to make ML accessible to a wider audience, including domain experts and non-researchers.
In this coding session, we focus on supervised classification with tabular data. You will learn the basics of Automated Machine Learning and how to use the Python-based AutoML tool Auto-sklearn developed at the Machine Learning Lab of the University of Freiburg.
We’ll do the exercises via Google-Colab, so you only need a laptop and an account. You can check your setup here: https://colab.research.google.com/drive/1-ToezX3hYomqPy4WqJLk2hvqoJzH5ji0. Also, we assume some experience with Python and basic ML knowledge.
9:30am - 12:30pm
Stefan-Meier-Straße 26, Institute of Medical Biometry and Statistics
Time | Plan |
---|---|
9:30-10:30 | Introduction |
10:30-12:00 | Hands-On Coding Session |
June 9th, 2022
9:30am – 5:00pm
HS Otto-Krayer-Haus
The short talks are 5-10 minutes long, and the intention is to introduce your work in an understandable way to those outside your immediate field. We encourage you to focus on the high-level approach you use so other researchers could get inspired by it regardless of the downstream applications. Remember that these short talks are an excellent opportunity to start some collaborations outside your lab!
Time | Plan |
---|---|
10:00-10:15 | Opening |
10:15-10:30 | Introduction by orga team |
10:30-11:30 | Imant Daunhawer and Thomas Sutter: Multimodal and Scalable VAEs |
11:30-13:00 | Break/Networking (at the venue) |
13:00-14:00 | Short Talks |
14:00-15:00 | Thomas Brox: Paths Towards Open World Regularization |
15:00-16:00 | Break/Networking |
16:00-17:00 | Louis-Philippe Morency: Multimodal AI: Understanding Human Behaviors |
17:00-Inf | Eating/Drinking together |
November 12th, 2021
9:30am – 5:00pm
HS 1199 in KG I
Prof. Dr. Anne-Laure Boulesteix (Ludwig Maximilian University of Munich)
Dr. Chris Rackauckas (Massachusetts Institute of Technology)
First prize winner: SimpleBits: Less Bits for more Interpretability
By Robin Schirrmeister
Second prize winner: Denosising and Segmentation Methods for Light and Electron Microscopy
By Joachim Greiner
Third prize winner: Learning the Optimal Analysis of Biomedical Data
By Clemens Kreutz
Time | Plan |
---|---|
09:30-9:45 | Opening speech: by Prof. Dr. Stefan Rensing |
9:45-10:15 | Opening speech: by Prof. Dr. Harald Binder |
10:15-10:30 | Introduction by orga team |
10:30-11:30 | Invited I: Anne-Laure Boulesteix (LMU) |
11:30-13:00 | Lunch break (at the venue) |
13:00-14:00 | Flash Talks |
14:00-14:30 | Coffee break (at the venue) |
14:30-15:30 | Invited II: Ruslan Salakhutdinov (CMU) |
15:30-16:30 | Invited III: Chris Rackauckas (MIT) |
16:30-17:00 | Future Outlook |
17:00-Inf | Dinner at “Blauer Fuchs” |
Our mission is to create, grow, and sustain a collaborative environment for Ph.D. and Postdoc researchers in the field of artificial intelligence. We want to become familiar with the expertise of different groups and possible opportunities to collaborate and help each other in a friendly environment. The important aim of this community is to increase the number of university-wide artificial intelligence-related events and share these possibilities with all active researchers in this field.
To be able to achieve our goals we are planning to encourage more members to be part of our community and hold workshops and science slams, invite keynote speakers, plan leisure time together to make the environment more friendly. As the first step, we have designed a short survey to know more about the preferences and how much time the members want to spend to participate in the community plans. After gathering this information, we are able to define some measurable operational goals.
If you want to be part of the organization team or get emails from us regarding nextgen_ai activities, please drop us an email.