This seminar simulates a machine learning conference, where the students take on the role of authors and reviewers. It consists of multiple phases.
Attend the mandatory first meeting either in person or remotely (details on TUWEL).
You select two topics/papers (i.e., two bullet points) from one of the topics below. You will work with the material mentioned in the overview and the topic-specific resources.
You choose your own topic to work on. This can be some existing machine learning paper/work or an own creative idea in the context of machine learning. We strongly encourage you to start from existing papers from the following venues: NeurIPS, ICML, ICLR, COLT, AISTATS, UAI, JMLR, MLJ. Importantly, your idea has to be specific and worked out well. Nevertheless, choose one of our suggestions as well.
Independent of the option you chose, understand the fundamentals of your topic and try to answer the following questions:
Select topics and write a short description of them together with the answers to the questions (~3 sentences should be sufficient) in TUWEL.
We can only accept your own proposals if you can answer the mentioned questions and have a well worked out topic.
You and your fellow students will act as reviewers and bid on the topics of your peers you want to review. Based on the biddings, we (in the role as chairs of the conference) will select one of each student’s proposals as the actual project you will work on for the rest of this semester. You do not need to work on the other project, anymore. Additionally, we will also assign two different projects from other students to you, which you will have to review later in the semester.
Now the actual work starts. Gather deep understanding of your topic, write a first draft of your report and give a 5-minute presentation. Feel free to go beyond the given material.
You will schedule two meetings with your supervisor to discuss your progress, but do not hesitate to contact him/her if you have any questions.
You will again act as a reviewer for the conference by writing two reviews, one for each draft report assigned to you.
Based on the reviews from your peers (and our feedback) you will further work on your topic.
Give a final presentation and submit your report.
You should have access to the literature and papers through Google scholar, DBLP, the provided links, or the TU library.
Motivation: Ability to learn logically definable concepts from labelled data is a theoretical model of Machine Learning which is explainable by design, and integrates ideas from both logic (especially finite model theory) and PAC Learning.
Overview:
Papers and topics:
Motivation
Full information about a many-body quantum system is usually out-of-reach due to the exponential growth – with the size of the system – of the number of parameters needed to encode its state. Nonetheless, in order to understand the complex phenomenology that can be observed in these systems, it is often sufficient to consider dynamical or stationary properties of local observables or, at most, of few-body correlation functions of a subsystem. One can formulate the problem of finding the generator of the subsystem dynamics as a variational problem, and solve them using the standard toolbox of machine learning for optimization.
Overview
H.-P. Breuer, F. Petruccione, The theory of open quantum systems Chapter 3.1 and 3.2
Papers and topics
Motivation: In active learning, the learning algorithm is allowed to select the data points it wants to see labelled, for example, where it is most uncertain. The goal is to reduce the labelling effort. This is useful in applications where unlabelled data is abundant, yet labels are scarce, such as node classification in social networks, drug discovery, and autonomous driving.
Overview:
Papers and topics:
Motivation: Classical algorithms cannot be combined with neural networks as training neural networks would require the computation of a gradient which is not possible for classical algorithms. Thus, we are interested in making classical algorithms differentiable.
Papers & Projects:
The above thesis gives an introduction to the field of differentiable algorithms and is built upon many conference publications by the author at strong ML conferences. We recommend picking either one of the chapters or one of the underlying publications from ICML, ICLR, NeurIPS or CVPR as a basis.
Motivation: Computing a disentangled representation is a very desirable property for modern deep learning architectures. Having access to individual, disentangled factors is expected to provide significant improvements for generalisation, interpretability and explainability.
Overview:
Papers and topics:
Motivation: Many datastructures have an innate structure that our neural networks should respect. For example the output of a graph neural networks should not change if we permute the vertices (permutation equivariance/invariance).
Overview:
Papers and topics:
Motivation: The ability of a model to adapt and perform well on new data is crucial. A model which generalises not only performs well on the training set, but on unseen data as well. Understanding and characterising why and how deep learning can generalise well is still an open question.
Overview:
Papers and topics:
Motivation: Graphs are a very general structure and can be applied to many areas: molecules and developing medicine, geographical maps, spread of diseases. They can be used to model physical systems and solve partial differential equations. Even images and text can be seen as a special case of graphs. Thus it makes sense to develop neural networks that can work with graphs. GNNs have strong connections to many classical computer science topics (algorithmics, logic, …) while also making use of neural networks. This means that work on GNN can be very theoretical, applied or anything in between.
Overview:
Papers & Projects:
Note: For very long papers we do not expect you to read the entire appendix.
Motivation: Large language models such as ChatGPT are seeing a huge research interest. Some companies are releasing more or less free models, and open-source initiatives have sprung up.
Overview: In this seminar paper, an overview of the latest large language models that are available in various forms is given. This includes, in particular, an investigation of their performance, and an explanation how performance can be evaluated objectively at all.
Advisor: Prof. Clemens Heitzinger
Motivation:
Large language models such as ChatGPT are seeing a huge research interest. Some companies are releasing more or less free models, and open-source initiatives have sprung up.
Overview:
In this seminar paper, an overview of the latest large language models that are available in various forms is given. This includes, in particular, an investigation of their performance, and an explanation how performance can be evaluated objectively at all.
Motivation: Exploration of new perspectives opened by ML methods for complex quantum systems and/or improving machine learning methods from the point of view as a physical system of interacting elements
Overview: Statistical field theory for neural networks (Lecture notes) by Moritz Helias and David Dahmen, https://arxiv.org/abs/1901.10416
Related Works:
Advisor: Prof. Sabine Andergassen
Task: Inferring Markovian quantum master equations of few-body observables in interacting spin chains
Motivation Full information about a many-body quantum system is usually out-of-reach due to the exponential growth – with the size of the system – of the number of parameters needed to encode its state. Nonetheless, in order to understand the complex phenomenology that can be observed in these systems, it is often sufficient to consider dynamical or stationary properties of local observables or, at most, of few-body correlation functions of a subsystem. One can formulate the problem of finding the generator of the subsystem dynamics as a variational problem, and solve them using the standard toolbox of machine learning for optimization.
Overview H.-P. Breuer, F. Petruccione, The theory of open quantum systems Chapter 3.1 and 3.2
Papers and topics
Advisor: Francesco Carnazza
Motivation:
Exploration of new perspectives opened by ML methods for complex quantum systems and/or improving machine learning methods from the point of view as a physical system of interacting elements
Overview:
Statistical field theory for neural networks (Lecture notes) by Moritz Helias and David Dahmen, https://arxiv.org/abs/1901.10416
Papers:
Motivation: Synthetic Aperture Radar (SAR) is an active microwave imaging system that provides high-resolution images day and night under all weather conditions. It has been widely used in many practical applications, such as environment, crop monitoring, and disaster detection. Using best-suited machine learning algorithms to derive useful information from these data is essential.
Overview:
Papers and topics:
Motivation: Transformers have revolutionized natural-language processing, and research has exploded since 2017.
Overview: In this seminar paper, the functioning of transformers is explained and an overview of the latest developments, regarding large language models, time-series predction, etc., is given.
Advisor: Prof. Clemens Heitzinger
Motivation:
Transformers have revolutionized natural-language processing, and research has exploded since 2017.
Overview:
In this seminar paper, the functioning of transformers is explained and an overview of the latest developments, regarding large language models, time-series predction, etc., is given.
Motivation: Machine learning systems are ubiquitous and it is necessary to make sure they behave as intended. In particular, trustworthiness can be achieved by means of privacy-preserving, robust, and explainable algorithms.
Overview:
Papers and topics: