Schedule

8:30 Registration
9:00 Opening & Welcome with Wilfried Gansterer
9:15 Dan Alistarh: Accurate Model Compression at GPT Scale
9:45 Talk Thomas Gärtner
10:15 Young Experts: Minute Madness
10:45 Coffee
11:15 Aristides Gionis: Submodular optimization and interpretable machine learning
11:45 Award Ceremony: OCG Förderpreis-FH, Heinz Zemanek Preis 2024
12:15 Lunch
14:00 Talk Sebastian Forster
14:30 Talk Fariba Karimi
15:00 New Professors Session
16:00 Coffee
16:30 Claudia Plant: Making Sense of Large Networks
17:00 Talk Jan Beutel
17:30 Closing

 Special Sessions

Young Experts: Minute Madness

In the "Young Experts: Minute Madness" session, excellent doctoral students in the field of computer science at Austrian universities will present their research work in a 1-minute overview talk, followed by a poster session.

New Professors Session

In the New Professors Session, recently hired professors in Austria will say "hi" and introduce themselves to the Austrian community. In short presentations, they provide insights into what drives them in their work and highlight their current and future research plans. We look forward to the following presentations:

 Speakers

Accurate Model Compression at GPT Scale

A key barrier to the wide deployment of highly-accurate machine learning models, whether for language or vision, is their high computational and memory overhead. Although we possess the mathematical tools for highly-accurate compression of such models, these theoretically-elegant techniques require second-order information of the model’s loss function, which is hard to even approximate efficiently at the scale of billion-parameter models. In this talk, I will describe our work on bridging this computational divide, which enables the accurate second-order pruning and quantization of models at truly massive scale. Compressed using our techniques, models with billions and even trillions of parameters can be executed efficiently on a few GPUs, with significant speedups, and negligible accuracy loss. Models created using our techniques have been downloaded millions of times from open-source repositories such as HuggingFace.

Dan Alistarh (Institute of Science and Technology Austria)

Dan Alistarh is a Professor at ISTA. Previously, he was a Visiting Professor at MIT, Researcher with Microsoft, a Postdoc at MIT CSAIL, and received his PhD from the EPFL. His research is on algorithms for efficient machine learning and high-performance computing, with a focus on scalable DNN inference and training, for which he was awarded ERC Starting and Proof-of-Concept Grants. In his spare time, he works with the ML research team at Neural Magic, a startup based in Boston, on making compression faster, more accurate and accessible to practitioners.


Submodular optimization and interpretable machine learning

Submodular functions are used to characterize the diminishing-returns property, which appears in many application areas, including information summarization, sensor placement, viral marketing, and more. Optimizing submodular functions has a rich history in mathematics and operations research, while recently, the subject has received increased attention due to the prevalent role of submodular functions in data-science applications. In this talk we will discuss two recent projects on the topic of interpretable classification, both of which make interesting connections with submodular optimization. For the first project, we address the problem of multi-label classification via concise and discriminative rule sets. Submodularity is used to account for diversity, which helps avoiding redundancy, and thus, controlling the number of rules in the solution set. In the second project we aim to find accurate decision trees that have small size, and thus, are interpretable. We study a general family of impurity functions, including the popular functions of entropy and Gini-index, and show that a simple enhancement, relying on the framework of adaptive submodular ranking, can be used to obtain a logarithmic approximation guarantee on the tree complexity.

Aristides Gionis (KTH Royal Institute of Technology)

Aristides Gionis is a WASP professor in KTH Royal Institute of Technology, Sweden, and an adjunct professor in Aalto University, Finland. He obtained his PhD from Stanford University, USA, and he has been a senior research scientist in Yahoo! Research. He has contributed in several areas of data science, such as data clustering and summarization, graph mining and social-network analysis, analysis of data streams, and privacy-preserving data mining. His current research is funded by the Wallenberg AI, Autonomous Systems and Software Program (WASP) and by the European Commission with an ERC Advanced grant (REBOUND) and the project SoBigData++.


Making Sense of Large Networks

A graph is a very simple concept, just consisting of nodes and edges. Real-world graphs exhibit a surprisingly rich set of patterns. For instance, nodes can form groups with different characteristics, e.g., groups of densely connected nodes, or groups of nodes that form bi-partite or tree-like patterns. Nodes can have different structural roles like hubs, spokes and bridges. Graphs are usually represented by adjacency matrices or edge lists. Even for a small graph, it is impossible to understand its patterns based on this representation. For large real-world social or biological networks, this is completely impossible. Therefore, we need methods for representation learning, clustering and summarization in order to make sense of networks. This talk covers some recent approaches, ranging from methods that follow an information-theoretic objective to deep representation learning approaches.

Claudia Plant (University of Vienna)

Claudia Plant is full professor, leader of the Data Mining and Machine Learning research group at the Faculty of Computer Science University of Vienna, Austria. Her group focuses on new methods for exploratory data mining, e.g., clustering, anomaly detection, graph mining and matrix factorization. Many approaches relate unsupervised learning to data compression, i.e. the better the found patterns compress the data the more information we have learned. Other methods rely on finding statistically independent patterns or multiple non-redundant solutions, on ensemble learning or on nature-inspired concepts such as synchronization. Indexing techniques and methods for parallel hardware support exploring massive data. Claudia Plant has co-authored over 150 peer-reviewed publications, among them more than 30 contributions to the top-level data mining conferences KDD and ICDM and 4 Best Paper Awards. Papers on scalability aspects appeared at SIGMOD, ICDE, and the results of interdisciplinary projects in leading application-related journals such as Bioinformatics, Cerebral Cortex and Water Research.


Jan Beutel received his MSc and PhD in Electrical Engineering from the Swiss Federal Institute of Technology (ETH), Zurich in 2000 and 2005 respectively. He has been with u-blox AG and spent time as a visiting researcher at the Berkeley Wireless Research Center. At ETH Zurich he has been heading a research group on networked embedded systems at the Computer Engineering and Networks Lab (TIK). In 2020, he joined the University of Innsbruck as a full Professor. In his research Jan Beutel has pioneered the use of in-situ wireless sensors for long-lived environmental monitoring and natural hazard mitigation applications, especially in high-mountain areas leading to many highly cited publications. The sensor networks and associated data on the slopes of the Matterhorn (CH) constitute the longest and densest data record in mountain permafrost research worldwide feeding into both basic research as well as international climate monitoring and policy making.


Sebastian Forster is a professor at the Department of Computer Science of the Paris Lodron University of Salzburg where he performs basic research in the areas of distributed and dynamic algorithms. Sebastian completed his PhD under the supervision of Monika Henzinger at the University of Vienna in 2015. His thesis on dynamic graph algorithms was awarded with the Heinz Zemanek Award of the Austrian Computer Society. He joined the Paris Lodron University of Salzburg in 2017 and received an ERC Starting Grant in 2020. Over the course of his career, he went on research stays at Microsoft Research in Mountain View (2014), the Simons Institute for the Theory of Computing at UC Berkeley (2015), the Max Planck Institute for Informatics in Saarbrücken (2016), and Google Research in Zurich (2023).

Thomas Gärtner has been Professor of Machine Learning at TU Wien since 2019. From 2015 to 2019 he had been full Professor of Data Science at the University of Nottingham. Before that, he had been leading a research group jointly hosted by the University of Bonn and Fraunhofer IAIS. During this time he received an award in the Emmy-Noether programme of the DFG. His main area of research is computationally efficient and effective machine learning algorithms with theoretical guarantees and practical demonstrations in real-world applications. He has been awarded grants from DFG, UKRI, FFG, WWTF, EU, and FWF. He gave tutorials at ICML and ECMLPKDD, was program co-chair of ECMLPKDD and of more than 10 international workshops. He has been an editor for the Machine Learning journal since 2006, a board member of the CAIML at TU Wien since 2021, and has regularly been Area Chair for ECMLPKDD, NeurIPS and/or ICML.

This list of speakers is still incomplete and will be finalized within the upcoming weeks.