Research Seminar

The IOL Seminar and Lecture Series at the Zuse Institute Berlin serves to bring together researchers presenting their latest work and to organize tutorial lectures on valuable topics not typically covered in graduate coursework. Presentations usually take place on Wednesday afternoons in ZIB's Seminar Room 2006.

For talks before year 2026, see 2024–2025, 2022–2023, and 2019–2021.

Upcoming talks

Photo Rebekka Burkholz Rebekka BurkholzCISPA Helmholtz Center for Information Security
@ ZIB, Room 0001 (Studio da Vinci)

Title. Towards AI That Is Smart, Sparse, and Social

Abstract.

Deep learning continues to achieve impressive breakthroughs across disciplines but relies on increasingly large neural network models that are trained on massive data sets. Their development inflicts costs that are only affordable by a few labs and prevent global participation in the creation of related technologies. But does it really have to be like this? We will identify some of the major challenges of deep learning at small scales and present solution strategies pertaining to the design of sparse training algorithms and problem specific neural network design in the context of agentic networks, which hold the promise to overcome a fundamental trade-off between model specialization and trainability.


Photo Karl Welzel Karl WelzelUniversity of Oxford
@ ZIB, Room 4027 (Roter Salon)

Title. Do third derivatives accelerate unconstrained continuous optimization?

Abstract.

Methods for unconstrained continuous optimization can be classified by the strength of the oracle for the objective function they require. First-order methods only require access to function values and gradients, second-order methods additionally require access to Hessians. Higher-order methods correspondingly require access to derivatives of order 1 to p for some p ≥ 3 and are called tensors methods. Intuitively, access to additional derivative information should speed up convergence to an approximate minimizer. We will cover the different perspectives from which this becomes true. First, convergence speed can be quantified from the perspective of global iteration complexity, for which the literature asserts that methods minimizing a regularized Taylor expansions, and in particular the adaptive ARp method, are optimal. Second, we will discuss local convergence results, showing how ARp can converge superlinearly even if the Hessian is singular at the minimizer. Third, numerical experiments on a range of test functions show that (after introducing certain heuristics) the third-order AR3 method needs fewer iterations and oracle calls than the second-order AR2 method.